%0 Generic %D 2023 %T Robust Decision-Focused Learning for Reward Transfer %A Abhishek Sharma %A Parbhoo, Sonali %A Omer Gottesman %A Finale Doshi-Velez %K Computer Science - Artificial Intelligence %K Computer Science - Machine Learning %X Decision-focused (DF) model-based reinforcement learning has recently been introduced as a powerful algorithm which can focus on learning the MDP dynamics which are most relevant for obtaining high rewards. While this approach increases the performance of agents by focusing the learning towards optimizing for the reward directly, it does so by learning less accurate dynamics (from a MLE standpoint), and may thus be brittle to changes in the reward function. In this work, we develop the robust decision-focused (RDF) algorithm which leverages the non-identifiability of DF solutions to learn models which maximize expected returns while simultaneously learning models which are robust to changes in the reward function. We demonstrate on a variety of toy example and healthcare simulators that RDF significantly increases the robustness of DF to changes in the reward function, without decreasing the overall return the agent obtains. %I arXiv %G eng %U http://arxiv.org/abs/2304.03365 %R 10.48550/arXiv.2304.03365 %0 Conference Paper %B The Eleventh International Conference on Learning Representations %D 2023 %T Performance Bounds for Model and Policy Transfer in Hidden-parameter MDPs %A Fu, Haotian %A Jiayu Yao %A Omer Gottesman %A Finale Doshi-Velez %A George Konidaris %X In the Hidden-Parameter MDP (HiP-MDP) framework, a family of reinforcement learning tasks is generated by varying hidden parameters specifying the dynamics and reward function for each individual task. HiP-MDP is a natural model for families of tasks in which meta- and lifelong-reinforcement learning approaches can succeed. Given a learned context encoder that infers the hidden parameters from previous experience, most existing algorithms fall into two categories: \$\textbackslashtextit\model transfer\\$ and \$\textbackslashtextit\policy transfer\\$, depending on which function the hidden parameters are used to parameterize. We characterize the robustness of model and policy transfer algorithms with respect to hidden parameter estimation error. We first show that the value function of HiP-MDPs is Lipschitz continuous under certain conditions. We then derive regret bounds for both settings through the lens of Lipschitz continuity. Finally, we empirically corroborate our theoretical analysis by experimentally varying the hyper-parameters governing the Lipschitz constants of two continuous control problems; the resulting performance is consistent with our predictions. %B The Eleventh International Conference on Learning Representations %G eng %U https://openreview.net/forum?id=20gBzEzgtiI %0 Journal Article %J Preprint %D 2023 %T Robust Decision-Focused Learning for Reward Transfer %A Abhishek Sharma %A Parbhoo, Sonali %A Omer Gottesman %A Finale Doshi-Velez %K Computer Science - Artificial Intelligence %K Computer Science - Machine Learning %X Decision-focused (DF) model-based reinforcement learning has recently been introduced as a powerful algorithm which can focus on learning the MDP dynamics which are most relevant for obtaining high rewards. While this approach increases the performance of agents by focusing the learning towards optimizing for the reward directly, it does so by learning less accurate dynamics (from a MLE standpoint), and may thus be brittle to changes in the reward function. In this work, we develop the robust decision-focused (RDF) algorithm which leverages the non-identifiability of DF solutions to learn models which maximize expected returns while simultaneously learning models which are robust to changes in the reward function. We demonstrate on a variety of toy example and healthcare simulators that RDF significantly increases the robustness of DF to changes in the reward function, without decreasing the overall return the agent obtains. %B Preprint %I arXiv %G eng %U http://arxiv.org/abs/2304.03365 %R 10.48550/arXiv.2304.03365 %0 Journal Article %J The 14th International Conference on Ambient Systems, Networks and Technologies Networks (ANT 2022) and The 6th International Conference on Emerging Data and Industry 4.0 (EDI40) %D 2023 %T Travel-time prediction using neural-network-based mixture models %A Abhishek Sharma %A Zhang, Jing %A Daniel Nikovski %A Finale Doshi-Velez %K mixture modeling %K Neural networks %K travel-time predictions %X Accurate estimation of travel times is an important step in smart transportation and smart building systems. Poor estimation of travel times results in both frustrated users and wasted resources. Current methods that estimate travel times usually only return point estimates, losing important distributional information necessary for accurate decision-making. We propose using neural network-based mixture distributions to predict a user's travel times given their origin and destination coordinates. We show that our method correctly estimates the travel time distribution, maximizes utility in a downstream elevator scheduling task, and is easy to retrain—making it a versatile and an inexpensive-to-maintain module when deployed in smart crowd management systems. %B The 14th International Conference on Ambient Systems, Networks and Technologies Networks (ANT 2022) and The 6th International Conference on Emerging Data and Industry 4.0 (EDI40) %V 220 %P 1033–1038 %G eng %U https://www.sciencedirect.com/science/article/pii/S1877050923006798 %R 10.1016/j.procs.2023.03.144 %0 Journal Article %J Preprint %D 2022 %T Addressing Leakage in Concept Bottleneck Models %A Havasi, Marton %A Parbhoo, Sonali %A Finale Doshi-Velez %B Preprint %8 2022 %G eng %U https://openreview.net/forum?id=tglniD_fn9 %0 Journal Article %J Preprint %D 2022 %T A Bayesian Approach to Learning Bandit Structure in Markov Decision Processes %A Zhang, Kelly W. %A Omer Gottesman %A Finale Doshi-Velez %X In the reinforcement learning literature, there are many algorithms developed for either Contextual Bandit (CB) or Markov Decision Processes (MDP) environments. However, when deploying reinforcement learning algorithms in the real world, even with domain expertise, it is often difficult to know whether it is appropriate to treat a sequential decision making problem as a CB or an MDP. In other words, do actions affect future states, or only the immediate rewards? Making the wrong assumption regarding the nature of the environment can lead to inefficient learning, or even prevent the algorithm from ever learning an optimal policy, even with infinite data. In this work we develop an online algorithm that uses a Bayesian hypothesis testing approach to learn the nature of the environment. Our algorithm allows practitioners to incorporate prior knowledge about whether the environment is that of a CB or an MDP, and effectively interpolate between classical CB and MDP-based algorithms to mitigate against the effects of misspecifying the environment. We perform simulations and demonstrate that in CB settings our algorithm achieves lower regret than MDP-based algorithms, while in non-bandit MDP settings our algorithm is able to learn the optimal policy, often achieving comparable regret to MDP-based algorithms. %B Preprint %I arXiv %8 jul %G eng %U http://arxiv.org/abs/2208.00250 %R 10.48550/arXiv.2208.00250 %0 Journal Article %J Proceedings of the AAAI Conference on Human Computation and Crowdsourcing %D 2022 %T Connecting Algorithmic Research and Usage Contexts: A Perspective of Contextualized Evaluation for Explainable AI %A Liao, Q. Vera %A Zhang, Yunfeng %A Luss, Ronny %A Finale Doshi-Velez %A Dhurandhar, Amit %X Recent years have seen a surge of interest in the field of explainable AI (XAI), with a plethora of algorithms proposed in the literature. However, a lack of consensus on how to evaluate XAI hinders the advancement of the field. We highlight that XAI is not a monolithic set of technologies–-researchers and practitioners have begun to leverage XAI algorithms to build XAI systems that serve different usage contexts, such as model debugging and decision-support. Algorithmic research of XAI, however, often does not account for these diverse downstream usage contexts, resulting in limited effectiveness or even unintended consequences for actual users, as well as difficulties for practitioners to make technical choices. We argue that one way to close the gap is to develop evaluation methods that account for different user requirements in these usage contexts. Towards this goal, we introduce a perspective of contextualized XAI evaluation by considering the relative importance of XAI evaluation criteria for prototypical usage contexts of XAI. To explore the context dependency of XAI evaluation criteria, we conduct two survey studies, one with XAI topical experts and another with crowd workers. Our results urge for responsible AI research with usage-informed evaluation practices, and provide a nuanced understanding of user requirements for XAI in different usage contexts. %B Proceedings of the AAAI Conference on Human Computation and Crowdsourcing %V 10 %P 147–159 %8 oct %G eng %U https://ojs.aaai.org/index.php/HCOMP/article/view/21995 %R 10.1609/hcomp.v10i1.21995 %0 Journal Article %J Algorithms %D 2022 %T Designing Reinforcement Learning Algorithms for Digital Interventions: Pre-Implementation Guidelines %A Anna L. Trella %A Zhang, Kelly W. %A Nahum-Shani, Inbal %A Shetty, Vivek %A Finale Doshi-Velez %A Murphy, Susan A. %X Online reinforcement learning (RL) algorithms are increasingly used to personalize digital interventions in the fields of mobile health and online education. Common challenges in designing and testing an RL algorithm in these settings include ensuring the RL algorithm can learn and run stably under real-time constraints, and accounting for the complexity of the environment, e.g., a lack of accurate mechanistic models for the user dynamics. To guide how one can tackle these challenges, we extend the PCS (predictability, computability, stability) framework, a data science framework that incorporates best practices from machine learning and statistics in supervised learning to the design of RL algorithms for the digital interventions setting. Furthermore, we provide guidelines on how to design simulation environments, a crucial tool for evaluating RL candidate algorithms using the PCS framework. We show how we used the PCS framework to design an RL algorithm for Oralytics, a mobile health study aiming to improve users’ tooth-brushing behaviors through the personalized delivery of intervention messages. Oralytics will go into the field in late 2022. %B Algorithms %V 15 %P 255 %G eng %U https://www.mdpi.com/1999-4893/15/8/255 %N 8 %R 10.3390/a15080255 %0 Journal Article %J Journal of Affective Disorders %D 2022 %T Do clinicians follow heuristics in prescribing antidepressants? %A Isaac Lage %A Melanie F. Pradier %A Thomas H. McCoy %A Roy H. Perlis %A Finale Doshi-Velez %X Background While clinicians commonly learn heuristics to guide antidepressant treatment selection, surveys suggest real-world prescribing practices vary widely. We aimed to determine the extent to which antidepressant prescriptions were consistent with commonly-advocated heuristics for treatment selection. Methods This retrospective longitudinal cohort study examined electronic health records from psychiatry and non-psychiatry practice networks affiliated with two large academic medical centers between March 2008 and December 2017. Patients included 45,955 individuals with a major depressive disorder or depressive disorder not otherwise specified diagnosis who were prescribed at least one of 11 common antidepressant medications. Specific clinical features that may impact prescribing choices were extracted from coded data, and analyzed for association with index prescription in logistic regression models adjusted for sociodemographic variables and provider type. Results Multiple clinical features yielded 10% or greater change in odds of prescribing, including overweight and underweight status and sexual dysfunction. These heuristics were generally applied similarly across hospital systems and psychiatrist and non-psychiatrist providers. Limitations These analyses rely on coded clinical data, which is likely to substantially underestimate prevalence of particular clinical features. Additionally, numerous other features that may impact prescribing choices are not able to be modeled. Conclusion Our results confirm the hypothesis that clinicians apply heuristics on the basis of clinical features to guide antidepressant prescribing, although the magnitude of these effects is modest, suggesting other patient- or clinician-level factors have larger effects. Funding This work was funded by NSF GRFP (grant no. DGE1745303), Harvard SEAS, the Center for Research on Computation and Society at Harvard, the Harvard Data Science Initiative, and a grant from the National Institute of Mental Health (grant no. 1R01MH106577). %B Journal of Affective Disorders %V 311 %P 110–114 %8 aug %G eng %U https://www.sciencedirect.com/science/article/pii/S0165032722004724 %R 10.1016/j.jad.2022.04.128 %0 Journal Article %J Journal of Affective Disorders %D 2022 %T Efficiently identifying individuals at high risk for treatment resistance in major depressive disorder using electronic health records %A Isaac Lage %A McCoy Jr, Thomas H. %A Roy H. Perlis %A Finale Doshi-Velez %K CV - Journals %X Background With the emergence of evidence-based treatments for treatment-resistant depression, strategies to identify individuals at greater risk for treatment resistance early in the course of illness could have clinical utility. We sought to develop and validate a model to predict treatment resistance in major depressive disorder using coded clinical data from the electronic health record. Methods We identified individuals from a large health system with a diagnosis of major depressive disorder receiving an index antidepressant prescription, and used a tree-based machine learning classifier to build a risk stratification model to identify those likely to experience treatment resistance. The resulting model was validated in a second health system. Results In the second health system, the extra trees model yielded an AUC of 0.652 (95% CI: 0.623–0.682); with sensitivity constrained at 0.80, specificity was 0.358 (95% CI: 0.300–0.413). Lift in the top quintile was 1.99 (95% CI: 1.76–2.22). Including additional data for the 4 weeks following treatment initiation did not meaningfully improve model performance. Limitations The extent to which these models generalize across additional health systems will require further investigation. Conclusion Electronic health records facilitated stratification of risk for treatment-resistant depression and demonstrated generalizability to a second health system. Efforts to improve upon such models using additional measures, and to understand their performance in real-world clinical settings, are warranted. %B Journal of Affective Disorders %V 306 %P 254–259 %G eng %U https://www.sciencedirect.com/science/article/pii/S0165032722001951 %R 10.1016/j.jad.2022.02.046 %0 Journal Article %J Preprint %D 2022 %T From Soft Trees to Hard Trees: Gains and Losses %A Zeng, Xin %A Jiayu Yao %A Finale Doshi-Velez %A Weiwei Pan %X Trees are widely used as interpretable models. However, when they are greedily trained they can yield suboptimal predictive performance. Training soft trees, with probabilistic splits rather than deterministic ones, provides a way to supposedly globally optimize tree models. For interpretability purposes, a hard tree can be obtained from a soft tree by binarizing the probabilistic splits, called hardening. Unfortunately, the good performance of the soft model is often lost after hardening. We systematically study two factors contributing to the performance drop: first, the loss surface of the soft tree loss has many local optima (and thus the logic for using the soft tree loss becomes less clear), and second, the relative values of the soft tree loss do not correspond to relative values of the hard tree loss. We also demonstrate that simple mitigation methods in literature do not fully mitigate the performance drop. %B Preprint %G eng %0 Journal Article %J Preprint %D 2022 %T Gathering Strength, Gathering Storms: The One Hundred Year Study on Artificial Intelligence (AI100) 2021 Study Panel Report %A Littman, Michael L. %A Ajunwa, Ifeoma %A Berger, Guy %A Craig Boutilier %A Currie, Morgan %A Finale Doshi-Velez %A Hadfield, Gillian %A Horowitz, Michael C. %A Isbell, Charles %A Kitano, Hiroaki %A Levy, Karen %A Lyons, Terah %A Mitchell, Melanie %A Shah, Julie %A Sloman, Steven %A Vallor, Shannon %A Toby Walsh %X In September 2021, the "One Hundred Year Study on Artificial Intelligence" project (AI100) issued the second report of its planned long-term periodic assessment of artificial intelligence (AI) and its impact on society. It was written by a panel of 17 study authors, each of whom is deeply rooted in AI research, chaired by Michael Littman of Brown University. The report, entitled "Gathering Strength, Gathering Storms," answers a set of 14 questions probing critical areas of AI development addressing the major risks and dangers of AI, its effects on society, its public perception and the future of the field. The report concludes that AI has made a major leap from the lab to people's lives in recent years, which increases the urgency to understand its potential negative effects. The questions were developed by the AI100 Standing Committee, chaired by Peter Stone of the University of Texas at Austin, consisting of a group of AI leaders with expertise in computer science, sociology, ethics, economics, and other disciplines. %B Preprint %I arXiv %G eng %U http://arxiv.org/abs/2210.15767 %R 10.48550/arXiv.2210.15767 %0 Journal Article %J Preprint %D 2022 %T Generalizing Off-Policy Evaluation From a Causal Perspective For Sequential Decision-Making %A Parbhoo, Sonali %A Joshi, Shalmali %A Finale Doshi-Velez %X Assessing the effects of a policy based on observational data from a different policy is a common problem across several high-stake decision-making domains, and several off-policy evaluation (OPE) techniques have been proposed. However, these methods largely formulate OPE as a problem disassociated from the process used to generate the data (i.e. structural assumptions in the form of a causal graph). We argue that explicitly highlighting this association has important implications on our understanding of the fundamental limits of OPE. First, this implies that current formulation of OPE corresponds to a narrow set of tasks, i.e. a specific causal estimand which is focused on prospective evaluation of policies over populations or sub-populations. Second, we demonstrate how this association motivates natural desiderata to consider a general set of causal estimands, particularly extending the role of OPE for counterfactual off-policy evaluation at the level of individuals of the population. A precise description of the causal estimand highlights which OPE estimands are identifiable from observational data under the stated generative assumptions. For those OPE estimands that are not identifiable, the causal perspective further highlights where more experimental data is necessary, and highlights situations where human expertise can aid identification and estimation. Furthermore, many formalisms of OPE overlook the role of uncertainty entirely in the estimation process.We demonstrate how specifically characterising the causal estimand highlights the different sources of uncertainty and when human expertise can naturally manage this uncertainty. We discuss each of these aspects as actionable desiderata for future OPE research at scale and in-line with practical utility. %B Preprint %I arXiv %G eng %U http://arxiv.org/abs/2201.08262 %R 10.48550/arXiv.2201.08262 %0 Journal Article %J Preprint %D 2022 %T Identification of Subgroups With Similar Benefits in Off-Policy Policy Evaluation %A Keramati, Ramtin %A Omer Gottesman %A Celi, Leo Anthony %B Preprint %G eng %0 Conference Paper %D 2022 %T Identifying Structure in the MIMIC ICU Dataset %A Chin, Zad %A Raval, Shivam %A Finale Doshi-Velez %A Wattenberg, Martin %A Celi, Leo Anthony %X The MIMIC-III dataset, containing trajectories of 40,000 ICU patients, is one of the most popular datasets in machine learning for health space. However, there has been very little systematic exploration to understand what is the natural structure of these data–-most analyses enforce some type of top-down clustering or embedding. We take a bottom-up approach, identifying consistent structures that are robust across a range of embedding choices. We identified two dominant structures sorted by either fraction-inspired oxygen or creatinine –- both of which were validated as the key features by our clinical co-author. Our bottom-up approach in studying the macro-structure of a dataset can also be adapted for other datasets. %I Proceedings of the Conference on Health, Inference, and Learning, 2022. %8 dec %G eng %U https://openreview.net/forum?id=3vfn-cmUYQF %0 Journal Article %J Proceedings of the AAAI Conference on Human Computation and Crowdsourcing %D 2022 %T “If it didn’t happen, why would I change my decision?”: How Judges Respond to Counterfactual Explanations for the Public Safety Assessment %A Yacoby, Yaniv %A Ben Green %A Jr, Christopher L. Griffin %A Finale Doshi-Velez %K CAREER %K IIS-1750358 %X Many researchers and policymakers have expressed excitement about algorithmic explanations enabling more fair and responsible decision-making. However, recent experimental studies have found that explanations do not always improve human use of algorithmic advice. In this study, we shed light on how people interpret and respond to counterfactual explanations (CFEs)–-explanations that show how a model's output would change with marginal changes to its input(s)–-in the context of pretrial risk assessment instruments (PRAIs). We ran think-aloud trials with eight sitting U.S. state court judges, providing them with recommendations from a PRAI that includes CFEs. We found that the CFEs did not alter the judges' decisions. At first, judges misinterpreted the counterfactuals as real–-rather than hypothetical–-changes to defendants. Once judges understood what the counterfactuals meant, they ignored them, stating their role is only to make decisions regarding the actual defendant in question. The judges also expressed a mix of reasons for ignoring or following the advice of the PRAI without CFEs. These results add to the literature detailing the unexpected ways in which people respond to algorithms and explanations. They also highlight new challenges associated with improving human-algorithm collaborations through explanations. %B Proceedings of the AAAI Conference on Human Computation and Crowdsourcing %V 10 %P 219–230 %8 oct %G eng %U https://ojs.aaai.org/index.php/HCOMP/article/view/22001 %R 10.1609/hcomp.v10i1.22001 %0 Journal Article %J npj Digital Medicine %D 2022 %T An interpretable RL framework for pre-deployment modeling in ICU hypotension management %A Zhang, Kristine %A Wang, Henry %A Du, Jianzhun %A Brian Chu %A Robles Arévalo, Aldo %A Kindle, Ryan %A Celi, Leo Anthony %A Finale Doshi-Velez %K CAREER %K IIS-1750358 %X Computational methods from reinforcement learning have shown promise in inferring treatment strategies for hypotension management and other clinical decision-making challenges. Unfortunately, the resulting models are often difficult for clinicians to interpret, making clinical inspection and validation of these computationally derived strategies challenging in advance of deployment. In this work, we develop a general framework for identifying succinct sets of clinical contexts in which clinicians make very different treatment choices, tracing the effects of those choices, and inferring a set of recommendations for those specific contexts. By focusing on these few key decision points, our framework produces succinct, interpretable treatment strategies that can each be easily visualized and verified by clinical experts. This interrogation process allows clinicians to leverage the model’s use of historical data in tandem with their own expertise to determine which recommendations are worth investigating further e.g. at the bedside. We demonstrate the value of this approach via application to hypotension management in the ICU, an area with critical implications for patient outcomes that lacks data-driven individualized treatment strategies; that said, our framework has broad implications on how to use computational methods to assist with decision-making challenges on a wide range of clinical domains. %B npj Digital Medicine %V 5 %P 1–10 %8 nov %G eng %U https://www.nature.com/articles/s41746-022-00708-4 %N 1 %R 10.1038/s41746-022-00708-4 %0 Conference Paper %B Proceedings of the Sixth Workshop on Structured Prediction for NLP %D 2022 %T A Joint Learning Approach for Semi-supervised Neural Topic Modeling %A Chiu, Jeffrey %A Mittal, Rajat %A Tumma, Neehal %A Abhishek Sharma %A Finale Doshi-Velez %K CAREER %K IIS-1750358 %K To Do: CV %X Topic models are some of the most popular ways to represent textual data in an interpret-able manner. Recently, advances in deep generative models, specifically auto-encoding variational Bayes (AEVB), have led to the introduction of unsupervised neural topic models, which leverage deep generative models as opposed to traditional statistics-based topic models. We extend upon these neural topic models by introducing the Label-Indexed Neural Topic Model (LI-NTM), which is, to the extent of our knowledge, the first effective upstream semi-supervised neural topic model. We find that LI-NTM outperforms existing neural topic models in document reconstruction benchmarks, with the most notable results in low labeled data regimes and for data-sets with informative labels; furthermore, our jointly learned classifier outperforms baseline classifiers in ablation studies. %B Proceedings of the Sixth Workshop on Structured Prediction for NLP %I Association for Computational Linguistics %C Dublin, Ireland %P 40–51 %8 may %G eng %U https://aclanthology.org/2022.spnlp-1.5 %R 10.18653/v1/2022.spnlp-1.5 %0 Conference Paper %B Proceedings of the 7th Machine Learning for Healthcare Conference %D 2022 %T Learning Optimal Summaries of Clinical Time-series with Concept Bottleneck Models %A Wu, Carissa %A Parbhoo, Sonali %A Havasi, Marton %A Finale Doshi-Velez %X Despite machine learning models’ state-of-the-art performance in numerous clinical prediction and intervention tasks, their complex black-box processes pose a great barrier to their real-world deployment. Clinical experts must be able to understand the reasons behind a model’s recommendation before taking action, as it is crucial to assess for criteria other than accuracy, such as trust, safety, fairness, and robustness. In this work, we enable human inspection of clinical timeseries prediction models by learning concepts, or groupings of features into high-level clinical ideas such as illness severity or kidney function. We also propose an optimization method which then selects the most important features within each concept, learning a collection of sparse prediction models that are sufficiently expressive for examination. On a real-world task of predicting vasopressor onset in ICU units, our algorithm achieves predictive performance comparable to state-of-the-art deep learning models while learning concise groupings conducive for clinical inspection. %B Proceedings of the 7th Machine Learning for Healthcare Conference %I PMLR %P 648–672 %G eng %U https://proceedings.mlr.press/v182/wu22a.html %0 Conference Paper %B Decision Awareness in Reinforcement Learning Workshop at ICML 2022 %D 2022 %T Leveraging Factored Action Spaces for Efficient Offline Reinforcement Learning in Healthcare %A Tang, Shengpu %A Maggie Makar %A Sjoding, Michael %A Finale Doshi-Velez %A Wiens, Jenna %K CV - Journals %X Many reinforcement learning (RL) applications have combinatorial action spaces, where each action is a composition of sub-actions. A standard RL approach ignores this inherent factorization structure, resulting in a potential failure to make meaningful inferences about rarely observed sub-action combinations; this is particularly problematic for offline settings, where data may be limited. In this work, we propose a form of linear Q-function decomposition induced by factored action spaces. We study the theoretical properties of our approach, identifying scenarios where it is guaranteed to lead to zero bias when used to approximate the Q-function. Outside the regimes with theoretical guarantees, we show that our approach can still be useful because it leads to better sample efficiency without necessarily sacrificing policy optimality, allowing us to achieve a better bias-variance trade-off. Across several offline RL problems using simulators and real-world datasets motivated by healthcare problems, we demonstrate that incorporating factored action spaces into value-based RL can result in better-performing policies. Our approach can help an agent make more accurate inferences within under-explored regions of the state-action space when applying RL to observational datasets. %B Decision Awareness in Reinforcement Learning Workshop at ICML 2022 %8 jul %G eng %U https://openreview.net/forum?id=wl_o_hilncS %0 Journal Article %J Journal of Machine Learning Research 23 (2022) %D 2022 %T Mitigating the Effects of Non-Identifiability on Inference for Bayesian Neural Networks with Latent Variables %A Yacoby, Yaniv %A Weiwei Pan %A Finale Doshi-Velez %X Bayesian Neural Networks with Latent Variables (BNN+LVs) capture predictive uncertainty by explicitly modeling model uncertainty (via priors on network weights) and environmental stochasticity (via a latent input noise variable). In this work, we first show that BNN+LV suffers from a serious form of non-identifiability: explanatory power can be transferred between the model parameters and latent variables while fitting the data equally well. We demonstrate that as a result, in the limit of infinite data, the posterior mode over the network weights and latent variables is asymptotically biased away from the ground-truth. Due to this asymptotic bias, traditional inference methods may in practice yield parameters that generalize poorly and misestimate uncertainty. Next, we develop a novel inference procedure that explicitly mitigates the effects of likelihood non-identifiability during training and yields high-quality predictions as well as uncertainty estimates. We demonstrate that our inference method improves upon benchmark methods across a range of synthetic and real data-sets. %B Journal of Machine Learning Research 23 (2022) %G eng %0 Journal Article %J Preprint %D 2022 %T Policy Optimization with Sparse Global Contrastive Explanations %A Jiayu Yao %A Parbhoo, Sonali %A Weiwei Pan %A Finale Doshi-Velez %X We develop a Reinforcement Learning (RL) framework for improving an existing behavior policy via sparse, user-interpretable changes. Our goal is to make minimal changes while gaining as much benefit as possible. We define a minimal change as having a sparse, global contrastive explanation between the original and proposed policy. We improve the current policy with the constraint of keeping that global contrastive explanation short. We demonstrate our framework with a discrete MDP and a continuous 2D navigation domain. %B Preprint %I arXiv %G eng %U http://arxiv.org/abs/2207.06269 %R 10.48550/arXiv.2207.06269 %0 Journal Article %J Preprint %D 2022 %T Success of Uncertainty-Aware Deep Models Depends on Data Manifold Geometry %A Penrod, Mark %A Termotto, Harrison %A Reddy, Varshini %A Jiayu Yao %A Finale Doshi-Velez %A Weiwei Pan %X For responsible decision making in safety-critical settings, machine learning models must effectively detect and process edge-case data. Although existing works show that predictive uncertainty is useful for these tasks, it is not evident from literature which uncertainty-aware models are best suited for a given dataset. Thus, we compare six uncertainty-aware deep learning models on a set of edge-case tasks: robustness to adversarial attacks as well as out-of-distribution and adversarial detection. We find that the geometry of the data sub-manifold is an important factor in determining the success of various models. Our finding suggests an interesting direction in the study of uncertainty-aware deep learning models. %B Preprint %I arXiv %8 aug %G eng %U http://arxiv.org/abs/2208.01705 %R 10.48550/arXiv.2208.01705 %0 Conference Proceedings %B Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society %D 2022 %T Towards Robust Off-Policy Evaluation via Human Inputs %A Singh, Harvineet %A Joshi, Shalmali %A Finale Doshi-Velez %A Himabindu Lakkaraju %B Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society %G eng %U https://dl.acm.org/doi/abs/10.1145/3514094.3534198?casa_token=QYwH8TIC-QYAAAAA:9bP2zidmphQac-O4CP0-uibVvKm-wFR4QiQNQsSKwTT20NGPnrncrxS_ZVoZOl3xnWcnrWopRpEH %0 Conference Paper %B Progress and Challenges in Building Trustworthy Embodied AI %D 2022 %T What Makes a Good Explanation?: A Harmonized View of Properties of Explanations %A Subhash, Varshini %A Chen, Zixi %A Havasi, Marton %A Weiwei Pan %A Finale Doshi-Velez %K CAREER %K Duplicate Presentation %K IIS-1750358 %X Interpretability provides a means for humans to verify aspects of machine learning (ML) models. Different tasks require explanations with different properties. However, presently, there is a lack of standardization in assessing properties of explanations: different papers use the same term to mean different quantities, and different terms to mean the same quantity. This lack of standardization prevents us from rigorously comparing explanation systems. In this work, we survey explanation properties defined in the current interpretable ML literature, we synthesize properties based on what they measure, and describe the trade-offs between different formulations of these properties. We provide a unifying framework for comparing properties of interpretable ML. %B Progress and Challenges in Building Trustworthy Embodied AI %8 nov %G eng %U https://openreview.net/forum?id=YDyLZWwpBK2 %0 Conference Paper %B Workshop on Trustworthy and Socially Responsible Machine Learning, NeurIPS 2022 %D 2022 %T What Makes a Good Explanation?: A Harmonized View of Properties of Explanations %A Chen, Zixi %A Subhash, Varshini %A Havasi, Marton %A Weiwei Pan %A Finale Doshi-Velez %K IIS-1750358 %X Interpretability provides a means for humans to verify aspects of machine learning (ML) models. Different tasks require explanations with different properties. However, presently, there is a lack of standardization in assessing properties of explanations: different papers use the same term to mean different quantities, and different terms to mean the same quantity. This lack of standardization prevents us from rigorously comparing explanation systems. In this work, we survey explanation properties defined in the current interpretable ML literature, we synthesize properties based on what they measure, and describe the trade-offs between different formulations of these properties. We provide a unifying framework for comparing properties of interpretable ML. %B Workshop on Trustworthy and Socially Responsible Machine Learning, NeurIPS 2022 %8 nov %G eng %U https://openreview.net/forum?id=TnFHizNosji %0 Journal Article %J Preprint %D 2022 %T (When) Are Contrastive Explanations of Reinforcement Learning Helpful? %A Narayanan, Sanjana %A Isaac Lage %A Finale Doshi-Velez %X Global explanations of a reinforcement learning (RL) agent's expected behavior can make it safer to deploy. However, such explanations are often difficult to understand because of the complicated nature of many RL policies. Effective human explanations are often contrastive, referencing a known contrast (policy) to reduce redundancy. At the same time, these explanations also require the additional effort of referencing that contrast when evaluating an explanation. We conduct a user study to understand whether and when contrastive explanations might be preferable to complete explanations that do not require referencing a contrast. We find that complete explanations are generally more effective when they are the same size or smaller than a contrastive explanation of the same policy, and no worse when they are larger. This suggests that contrastive explanations are not sufficient to solve the problem of effectively explaining reinforcement learning policies, and require additional careful study for use in this context. %B Preprint %I arXiv %G eng %U http://arxiv.org/abs/2211.07719 %R 10.48550/arXiv.2211.07719 %0 Conference Paper %B Proceedings of The 25th International Conference on Artificial Intelligence and Statistics %D 2022 %T Wide Mean-Field Bayesian Neural Networks Ignore the Data %A Beau Coker %A Bruinsma, Wessel P. %A Burt, David R. %A Weiwei Pan %A Finale Doshi-Velez %X Bayesian neural networks (BNNs) combine the expressive power of deep learning with the advantages of Bayesian formalism. In recent years, the analysis of wide, deep BNNs has provided theoretical insight into their priors and posteriors. However, we have no analogous insight into their posteriors under approximate inference. In this work, we show that mean-field variational inference entirely fails to model the data when the network width is large and the activation function is odd. Specifically, for fully-connected BNNs with odd activation functions and a homoscedastic Gaussian likelihood, we show that the optimal mean-field variational posterior predictive (i.e., function space) distribution converges to the prior predictive distribution as the width tends to infinity. We generalize aspects of this result to other likelihoods. Our theoretical results are suggestive of underfitting behavior previously observered in BNNs. While our convergence bounds are non-asymptotic and constants in our analysis can be computed, they are currently too loose to be applicable in standard training regimes. Finally, we show that the optimal approximate posterior need not tend to the prior if the activation function is not odd, showing that our statements cannot be generalized arbitrarily. %B Proceedings of The 25th International Conference on Artificial Intelligence and Statistics %I PMLR %P 5276–5333 %G eng %U https://proceedings.mlr.press/v151/coker22a.html %0 Conference Proceedings %B proceeding at the Human Centered Explainable AI Conference: CHI Workshop on Human Centered Explainable AI (HCXAI %D 2022 %T "If it didn’t happen, why would I change my decision?": How Judges Respond to Counterfactual Explanations for the Public Safety Assessment" %A Yacoby, Y. %A B. Green %A C. Griffin, Jr. %A Doshi-Velez, F. %B proceeding at the Human Centered Explainable AI Conference: CHI Workshop on Human Centered Explainable AI (HCXAI %V 2 %P 1-24 %G eng %0 Conference Paper %B Proceedings of the Conference on Health, Inference, and Learning %D 2022 %T Identification of Subgroups With Similar Benefits in Off-Policy Policy Evaluation %A R. Keramati %A Gottesman, O. %A Celi, L. %A Doshi-Velez, F. %A Brunskill, E. %B Proceedings of the Conference on Health, Inference, and Learning %P 397-410 %G eng %0 Conference Paper %B Advances in Neural Information Processing Systems %D 2021 %T Learning MDPs from Features: Predict-Then-Optimize for Sequential Decision Making by Reinforcement Learning %A Wang, Kai %A Shah, Sanket %A Chen, Haipeng %A Perrault, Andrew %A Finale Doshi-Velez %A Milind Tambe %K MURI Grant Number W911NF-17-1-0370 %K MURI Grant Number W911NF-18-1-0208 %K NSF CAREER IIS-1750358 %X In the predict-then-optimize framework, the objective is to train a predictive model, mapping from environment features to parameters of an optimization problem, which maximizes decision quality when the optimization is subsequently solved. Recent work on decision-focused learning shows that embedding the optimization problem in the training pipeline can improve decision quality and help generalize better to unseen tasks compared to relying on an intermediate loss function for evaluating prediction quality. We study the predict-then-optimize framework in the context of sequential decision problems (formulated as MDPs) that are solved via reinforcement learning. In particular, we are given environment features and a set of trajectories from training MDPs, which we use to train a predictive model that generalizes to unseen test MDPs without trajectories. Two significant computational challenges arise in applying decision-focused learning to MDPs: (i) large state and action spaces make it infeasible for existing techniques to differentiate through MDP problems, and (ii) the high-dimensional policy space, as parameterized by a neural network, makes differentiating through a policy expensive. We resolve the first challenge by sampling provably unbiased derivatives to approximate and differentiate through optimality conditions, and the second challenge by using a low-rank approximation to the high-dimensional sample-based derivatives. We implement both Bellman-based and policy gradient-based decision-focused learning on three different MDP problems with missing parameters, and show that decision-focused learning performs better in generalization to unseen tasks. %B Advances in Neural Information Processing Systems %I Curran Associates, Inc. %V 34 %P 8795–8806 %G eng %U https://proceedings.neurips.cc/paper/2021/hash/49e863b146f3b5470ee222ee84669b1c-Abstract.html %0 Conference Paper %B proceeding at the International Conference on Machine Learning %D 2021 %T State Relevance for Off-Policy Evaluation %A S. Shein %A Ma, Y. %A Gottesman, O. %A Doshi-Velez, F. %B proceeding at the International Conference on Machine Learning %V 1 %P 1-20 %G eng %U arXiv:2109.06310v1 %0 Conference Paper %B proceedings at the International Conference on Machine Learning for Healthcare %D 2021 %T Power Constrained Bandits %A Futoma, J. %A M. Simons %A Doshi-Velez, F. %A Kamaleswaran, R. %B proceedings at the International Conference on Machine Learning for Healthcare %V 4 %P 1-50 %G eng %0 Conference Proceedings %B proceeding at the Clinical Research Informatics American Medical Informatics Association Summit (AMIA), %D 2021 %T Preferential Mixture-of-Experts: Interpretable Models that Rely on Human Expertise as Much as Possible %A Pradier, M. %A J. Zazo %A S. Parbhoo %A Perlis, R. %A Zazzi, M. %A Doshi-Velez, F. %B proceeding at the Clinical Research Informatics American Medical Informatics Association Summit (AMIA), %V 1 %P 1-15 %G eng %0 Journal Article %J Translational psychiatry %D 2021 %T How machine-learning recommendations influence clinician treatment selections: the example of antidepressant selection %A Jacobs, M. %A Pradier, M. %A McCoy, T. %A Perlis, R. %A Doshi-Velez, F. %A Gajos, K. %B Translational psychiatry %V 1 %P 1-9 %G eng %0 Conference Proceedings %B Proceeding at the International Conference on Machine Learning (ICML) %D 2021 %T Benchmarks, Algorithms, and Metrics for Hierarchical Disentanglement %A Ross, A. %A Doshi-Velez, F. %B Proceeding at the International Conference on Machine Learning (ICML) %V 2 %P 1-23 %G eng %0 Journal Article %J Computational Linguistics %D 2021 %T Depth-Bounded Statistical PCFG Induction as a Model of Human Grammar Acquisition %A L. Jin %A L. Schwartz %A Doshi-Velez, F. %A Miller, T. %A W. Schuler %B Computational Linguistics %V 47 %P 1-36 %G eng %N 1 %0 Journal Article %J Preprint %D 2021 %T Learning Under Adversarial and Interventional Shifts %A H. Singh %A S. Joshi %A Doshi-Velez, F. %A Lakkaraju, H. %B Preprint %V 1 %P 1-19 %G eng %0 Journal Article %J AI Magazine %D 2021 %T Machine Learning Techniques for Accountability %A B. Kim %A Doshi-Velez, F. %B AI Magazine %V 42 %P 1 %G eng %N 1 %0 Conference Proceedings %B proceedings at the Conference on Human Factors in Computing Systems (CHI) %D 2021 %T Designing AI for Trust and Collaboration in Time-Constrained Medical Decisions: A Sociotechnical Lens %A Jacobs, M. %A J. He %A M F. Pradier %A B. Lam %A A. Ahn %A McCoy, T. %A Perlis, R. %A Doshi-Velez, F. %A Gajos, K. %B proceedings at the Conference on Human Factors in Computing Systems (CHI) %V 2 %P 1-14 %G eng %0 Conference Proceedings %B proceedings at the Conference on Human Factors in Computing Systems (CHI) %D 2021 %T Designing AI for Trust and Collaboration in Time-Constrained Medical Decisions: A Sociotechnical Lens %A Jacobs, M. %A J. He %A M F. Pradier %A B. Lam %A A. Ahn %A McCoy, T. %A Perlis, R. %A Doshi-Velez, F. %A Gajos, K. %B proceedings at the Conference on Human Factors in Computing Systems (CHI) %V 2 %P 1-14 %G eng %0 Journal Article %J proceedings at the Machine Learning for Healthcare Conference %D 2021 %T Power Constrained Bandit %A Yao, J. %A Brunskill, E. %A Pan, W. %A S. Murphy %A Doshi-Velez, F. %B proceedings at the Machine Learning for Healthcare Conference %V 4 %P 1-50 %G eng %U https://static1.squarespace.com/static/59d5ac1780bd5ef9c396eda6/t/60fb3a37b5d5c73c4aba9c58/1627077177096/Power_Constrained_Bandits_MLHC_2021.pdf %0 Journal Article %J Critical Care Explorations %D 2021 %T Generalization in Clinical Prediction Models: The Blessing and Curse of Measurement Indicator Variables %A Futoma, J. %A M. Simons %A Doshi-Velez, F. %A Kamaleswaran, R. %B Critical Care Explorations %V 1 %P 1-11 %G eng %0 Conference Proceedings %B proceeding at the International Conference on Machine Learning: Information-Theoretic Methods for Rigorous, Responsible, and Reliable Machine Learning (ICML), 2021 %D 2021 %T Prediction-focused Mixture Models %A Narayanan, S. %A Sharma, A. %A Zeng, C. %A Doshi-Velez, F. %B proceeding at the International Conference on Machine Learning: Information-Theoretic Methods for Rigorous, Responsible, and Reliable Machine Learning (ICML), 2021 %V 1 %P 1-9 %G eng %0 Conference Proceedings %B proceedings at the International Conference on Machine Learning: Workshop on Uncertainty & Robustness in Deep Learning (ICML) %D 2021 %T Wide Mean-Field Variational Bayesian Neural Networks Ignore the Data %A Coker, B. %A S. Parbhoo %A Doshi-Velez, F. %B proceedings at the International Conference on Machine Learning: Workshop on Uncertainty & Robustness in Deep Learning (ICML) %V 1 %P 1-15 %G eng %0 Conference Proceedings %B proceedings at the International Conference on Machine Learning: Workshop on Uncertainty & Robustness in Deep Learning (ICML) %D 2021 %T Wide Mean-Field Variational Bayesian Neural Networks Ignore the Data %A Coker, B. %A S. Parbhoo %A Doshi-Velez, F. %B proceedings at the International Conference on Machine Learning: Workshop on Uncertainty & Robustness in Deep Learning (ICML) %V 1 %P 1-15 %G eng %0 Conference Proceedings %B proceeding at the International Conference on Machine Learning: Workshop on Theoretic Foundation, Criticism, and Application Trend of Explainable AI, %D 2021 %T Promises and Pitfalls of Black-Box Concept Learning Models %A A. Mahinpei %A J. Clark %A I. Lage %A Doshi-Velez, F. %A P. WeiWei %B proceeding at the International Conference on Machine Learning: Workshop on Theoretic Foundation, Criticism, and Application Trend of Explainable AI, %V 1 %P 1-13 %G eng %0 Journal Article %J Journal of AI Research (JAIR) %D 2021 %T Optimizing for Interpretability in Deep Neural Networks with Tree Regularization %A Wu, M. %A S. Parbhoo %A Hughes, M. %A V. Roth %A Doshi-Velez, F. %B Journal of AI Research (JAIR) %V 1 %P 1-37 %G eng %0 Conference Proceedings %B proceeding at the International Conference on Machine Learning: Workshop on Neglected Assumptions in Causal Inference (ICML) %D 2021 %T Pre-emptive Learning to Defer for Sequential Medical Decision-Making Under Uncertainty %A S. Parbhoo %A S. Shalmali %A Doshi-Velez, F. %B proceeding at the International Conference on Machine Learning: Workshop on Neglected Assumptions in Causal Inference (ICML) %V 1 %P 1-13 %G eng %0 Conference Proceedings %B proceeding at the International Conference on Machine Learning: Workshop on Neglected Assumptions in Causal Inference (ICML) %D 2021 %T On formalizing causal off-policy evaluation for sequential decision-making %A Parbhoo, S %A Shalmali, J %A Doshi-Velez, F. %B proceeding at the International Conference on Machine Learning: Workshop on Neglected Assumptions in Causal Inference (ICML) %G eng %0 Conference Proceedings %B proceeding at the Conference on American Medical Informatics Association (AMIA) %D 2021 %T Learning Predictive and Interpretable Timeseries Summaries from ICU Data %A Johnson, N %A Parbhoo, S %A Ross, A %A Doshi-Velez, F. %B proceeding at the Conference on American Medical Informatics Association (AMIA) %V 1 %P 1-10 %G eng %0 Conference Proceedings %B proceeding at the Conference on American Medical Informatics Association (AMIA) %D 2021 %T Learning Predictive and Interpretable Timeseries Summaries from ICU Data %A Johnson, N %A Parbhoo, S %A Ross, A %A Doshi-Velez, F. %B proceeding at the Conference on American Medical Informatics Association (AMIA) %V 1 %P 1-10 %G eng %0 Journal Article %J Translational Psychiatry %D 2021 %T How machine learning recommendations influence clinician treatment selections: example of antidepressant selection %A Jacobs, M. %A Pradier, M. %A McCoy, T. %A P. Roy %A Doshi-Velez, F. %A G. Krzysztof %B Translational Psychiatry %V 1 %P 1-9 %G eng %0 Conference Paper %B proceeding at the Conference on Human Factors in Computing Systems (CHI), 2021 %D 2021 %T Evaluating the Interpretability of Generative Models by Interactive Reconstruction %A Ross, A. %A N. Chen %A E. Hang %A E. Glassman %A Doshi-Velez, F. %B proceeding at the Conference on Human Factors in Computing Systems (CHI), 2021 %V 1 %P 1-20 %G eng %0 Conference Paper %B proceeding at the Conference on Human Factors in Computing Systems (CHI) %D 2021 %T Designing AI for Trust in Collaboration in Time-Constrained Medical Decisions: A Sociotechnical Lens %A Jacobs, M. %A J. He %A Pradier, M. %A B. Lam %A A. Ahn %A McCoy, T. %A Perlis, R. %A Doshi-Velez, F. %A Gajos, K. %B proceeding at the Conference on Human Factors in Computing Systems (CHI) %V 1 %P 1-14 %G eng %0 Conference Paper %B NeurIPS Workshop on Machine Learning for Health %D 2021 %T Identifying Decision Points for Safe and Interpretable Reinforcement Learning in Hypotension Treatment %A Zhang, K. %A Wang, H. %A Du, J. %A B. Chu %A Kindle, R. %A Celi, L. %A Doshi-Velez, F. %B NeurIPS Workshop on Machine Learning for Health %V 1 %P 1-9 %G eng %0 Conference Paper %B American Medical Informatics Association (AMIA) %D 2021 %T Preferential Mixture-of-Experts: Interpretable Models that Rely on Human Expertise as Much as Possible %A Pradier, M. %A J. Zazo %A S. Parbhoo %A Perlis, R. %A Zazzi, M. %A Doshi-Velez, F. %B American Medical Informatics Association (AMIA) %V 1 %P 1-13 %G eng %0 Journal Article %J The Lancet Digital Health %D 2020 %T The myth of generalisability in clinical research and machine learning in health care %A Futoma, J. %A M. Simons %A T. Panch %A Doshi-Velez, F. %A Celi, L. %B The Lancet Digital Health %V 3 %P 1-19 %G eng %0 Conference Proceedings %B presented at the International Conference on Machine Learning %D 2020 %T Interpretable Off-Policy Evaluation in Reinforcement Learning by Highlighting Influential Transitions %A Gottesman, O. %A Futoma, J. %A Liu, Y. %A S. Parbhoo %A Celi, L. %A Brunskill, E. %A Doshi-Velez, F. %B presented at the International Conference on Machine Learning %V 3 %P 1-19 %G eng %0 Journal Article %J presented at the International Conference on Machine Learning: Workshop on Human Interpretability in Machine Learnin %D 2020 %T Learning Interpretable Concept-Based Models with Human Feedback %A I. Lage %A Doshi-Velez, F. %B presented at the International Conference on Machine Learning: Workshop on Human Interpretability in Machine Learnin %V 1 %P 1-11 %G eng %0 Journal Article %J presented at the International Conference on Machine Learning: Workshop on Human Interpretability in Machine Learnin %D 2020 %T Learning Interpretable Concept-Based Models with Human Feedback %A I. Lage %A Doshi-Velez, F. %B presented at the International Conference on Machine Learning: Workshop on Human Interpretability in Machine Learnin %V 1 %P 1-11 %G eng %0 Journal Article %J Preprint %D 2020 %T Artificial Intelligence & Cooperation %A E. Bertino %A Doshi-Velez, F. %A M. Gini %A D. Lopresti %A D. Parkes %B Preprint %V 1 %P 1-4 %G eng %0 Journal Article %J Neuropsychopharmacology %D 2020 %T Predicting change in diagnosis from major depression to bipolar disorder after antidepressant initiation %A Pradier, M. %A C. Hughes %A McCoy, T. %A S. Barroilhet %A Doshi-Velez, F. %A Perlis, R. %B Neuropsychopharmacology %V 1 %P 1-7 %G eng %0 Journal Article %J Pediatrics %D 2020 %T Big Data in the Assessment of Pediatric Medication Safety %A McMahon, A. %A Cooper, W. %A J. Brown %A Carleton, B. %A Doshi-Velez, F. %A Kohane, I. %A Goldman, J. %A Hoffman, M. %A Kamaleswaran, R. %A Sakiyama, M. %A Sekine, S. %A Sturkenboom, M. %A Turner, M. %A Califf, R. %B Pediatrics %V 1 %P 1-11 %G eng %U https://finale.seas.harvard.edu/files/finale/files/big_data_in_the_assessment_of_pediatric_medication_safety.pdf %0 Journal Article %J Pediatrics %D 2020 %T Big Data in the Assessment of Pediatric Medication Safety %A McMahon, A. %A Cooper, W. %A J. Brown %A Carleton, B. %A Doshi-Velez, F. %A Kohane, I. %A Goldman, J. %A Hoffman, M. %A Kamaleswaran, R. %A Sakiyama, M. %A Sekine, S. %A Sturkenboom, M. %A Turner, M. %A Califf, R. %B Pediatrics %V 1 %P 1-11 %G eng %U https://finale.seas.harvard.edu/files/finale/files/big_data_in_the_assessment_of_pediatric_medication_safety.pdf %0 Conference Paper %B proceeding at the Conference on Neural Information Processing Systems (NeurIPS): Workshop on Real World Reinforcement Learning %D 2020 %T A Bayesian Approach to Learning Bandit Structure in Markov Decision Processes %A Zhang, K. %A Gottesman, O. %A Doshi-Velez, F. %B proceeding at the Conference on Neural Information Processing Systems (NeurIPS): Workshop on Real World Reinforcement Learning %P 1-12 %G eng %0 Conference Paper %B proceedings at the Conference on Neural Information Processing Systems (NeurIPS): Workshop on Real World Reinforcement Learning %D 2020 %T A Bayesian Approach to Learning Bandit Structure in Markov Decision Processes %A Zhang, K. %A Gottesman, O. %A Doshi-Velez, F. %B proceedings at the Conference on Neural Information Processing Systems (NeurIPS): Workshop on Real World Reinforcement Learning %P 1-12 %G eng %0 Report %D 2020 %T Artificial Intelligence and Cooperation %A E. Bertino %A Doshi-Velez, F. %A M. Gini %A D. Lopresti %A D. Parkes %7 1 %P 1-4 %G eng %U arXiv:2012.06034 %0 Journal Article %J JAMA Network Open %D 2020 %T Assessment of a Prediction Model for Antidepressant Treatment Stability Using Supervised Topic Models %A Hughes, M. %A Pradier, M. %A Ross, A. %A M. McCoy %A Perlis, R. %A Doshi-Velez, F. %B JAMA Network Open %P 1-14 %G eng %0 Journal Article %J proceeding at the Conference on Neural Information Processing Systems (NeurIPS) %D 2020 %T Incorporating Interpretable Output Constraints in Bayesian Neural Networks %A W. Yang %A Lorch, L. %A Graule, M. %A Lakkaraju, H. %A Doshi-Velez, F. %B proceeding at the Conference on Neural Information Processing Systems (NeurIPS) %V 2 %P 1-17 %G eng %0 Journal Article %J Neuropsychopharmacology %D 2020 %T Predicting change in diagnosis from major depression to bipolar disorder after antidepressant initiation %A Pradier, M. %A Hughes, M. %A McCoy, T. %A S. Barroilhet %A Doshi-Velez, F. %A Perlis, R. %B Neuropsychopharmacology %V 1 %P 1-7 %G eng %0 Conference Paper %B NeurIPS Workshop on Offline Reinforcement Learning %D 2020 %T Shaping Control Variates for Off-Policy Evaluation %A S. Parbhoo %A Gottesman, O. %A Doshi-Velez, F. %B NeurIPS Workshop on Offline Reinforcement Learning %P 1-9 %G eng %0 Journal Article %J Lancet Digital Health %D 2020 %T The Myth of Generalizability in Clinical Research and Machine Learning in Healthcare %A Futoma, J. %A M. Simons %A T. Panch %A Doshi-Velez, F. %A Celi, L. %B Lancet Digital Health %V 2 %P 1-5 %G eng %U https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7444947/ %0 Conference Paper %B proceeding at the Machine Learning for Healthcare Conference %D 2020 %T Transfer Learning from Well-Curated to Less-Resourced Populations with HIV %A S. Parbhoo %A M. Wieser %A V. Roth %A Doshi-Velez, F. %B proceeding at the Machine Learning for Healthcare Conference %P 1-20 %G eng %0 Conference Paper %B Neural Information Processing Systems Conference (NeurIPS) 2020 %D 2020 %T Model-Based Reinforcement Learning for Semi-Markov Decision Processes with Neural ODEs. %A Du, J. %A Futoma, J. %A Doshi-Velez, F. %B Neural Information Processing Systems Conference (NeurIPS) 2020 %V 2 %P 1-21 %G eng %0 Journal Article %J ICML Workshop on Human Interpretability in Machine Learning, %D 2020 %T Human-in-the-Loop Learning of Interpretable and Intuitive Representations %A I. Lage %A Doshi-Velez, F. %B ICML Workshop on Human Interpretability in Machine Learning, %V 1 %P 1-10 %G eng %0 Journal Article %J ICML Workshop on Theoretical Foundations of Reinforcement Learning %D 2020 %T Power-Constrained Bandits %A Yao, J. %A Brunskill, E. %A Pan, W. %A S. Murphy %A Doshi-Velez, F. %B ICML Workshop on Theoretical Foundations of Reinforcement Learning %V 2 %P 1-30 %G eng %0 Journal Article %J UAI %D 2020 %T PoRB-Nets: Poisson Process Radial Basis Function Networks %A Coker, B. %A M. Fernandez-Pradier %A Doshi-Velez, F. %B UAI %P 1-59 %G eng %0 Journal Article %J ICML Workshop on Theoretical Foundations of Reinforcement Learning %D 2020 %T PAC Imitation and Model-based Batch Learning of Contextual MDPs %A Y. Nair %A Doshi-Velez, F. %B ICML Workshop on Theoretical Foundations of Reinforcement Learning %V 2 %P 1-21 %G eng %0 Journal Article %J ICML Workshop on Inductive Biases, Invariances and Generalization in RL %D 2020 %T PAC Imitation and Model-based Batch Learning of Contextual MDPs %A Y. Nair %A Doshi-Velez, F. %B ICML Workshop on Inductive Biases, Invariances and Generalization in RL %V 2 %P 1-21 %G eng %0 Journal Article %J ICML Workshop on Uncertainty in Deep Learning %D 2020 %T Learned Uncertainty-Aware (LUNA) Bases for Bayesian Regression using Multi-Headed Auxiliary Networks %A S. Thakur %A C. Lorsung %A Yacoby, Y. %A Doshi-Velez, F. %A Pan, W. %B ICML Workshop on Uncertainty in Deep Learning %V 2 %P 1-18 %G eng %0 Journal Article %J AMIA %D 2020 %T Is Deep Reinforcement Learning Ready for Practical Applications in Healthcare? A Sensitivity Analysis of Duel-DDQN for Sepsis Treatment %A M. Lu %A Z. Shahn %A D. Sow %A Doshi-Velez, F. %A L. Lehman %B AMIA %V 1 %P 1-13 %G eng %0 Conference Paper %B International Conference on Machine Learning %D 2020 %T Interpretable Off-Policy Evaluation in Reinforcement Learning by Highlighting Influential Transitions %A Gottesman, O. %A Futoma, J. %A Liu, Y. %A S. Parbhoo %A LA. Celi %A Brunskill, E. %A Doshi-Velez, F. %B International Conference on Machine Learning %V 2 %P 1-17 %G eng %0 Journal Article %J International Conference on Machine Learning (IMCL). %D 2020 %T Interpretable Off-Policy Evaluation in Reinforcement Learning by Highlighting Influential Transitions %A Gottesman, O. %A Futoma, J. %A Liu, Y. %A S. Parbhoo %A LA. Celi %A Brunskill, E. %A Doshi-Velez, F. %B International Conference on Machine Learning (IMCL). %V 2 %P 1-17 %G eng %0 Journal Article %J ICML Workshop on Uncertainty in Deep Learning %D 2020 %T Failures of Variational Autoencoders and their Effects on Downstream Tasks %A Yacoby, Y. %A Pan, W. %A Doshi-Velez, F. %B ICML Workshop on Uncertainty in Deep Learning %V 1 %P 1-39 %G eng %0 Journal Article %J International Statistical Review %D 2020 %T Discussions on Horseshoe Regularisation for Machine Learning in Complex and Deep Models %A Ghosh, S. %A Doshi-Velez, F. %B International Statistical Review %V 1 %P 1-3 %G eng %0 Journal Article %J ICML Workshop on Human Interpretability in Machine Learning %D 2020 %T CRUDS: Counterfactual Recourse Using Disentangled Subspaces %A M. Downs, %A J. Chu, %A Yacoby, Y. %A Doshi-Velez, F. %A P. WeiWei %B ICML Workshop on Human Interpretability in Machine Learning %P 1-23 %G eng %0 Journal Article %J ICML Workshop on Uncertainty in Deep Learning %D 2020 %T BaCOUn: Bayesian Classifers with Out-of-Distribution Uncertainty %A T. Guenais %A D. Vamvourellis %A Yacoby, Y. %A Doshi-Velez, F. %A Pan, W. %B ICML Workshop on Uncertainty in Deep Learning %V 1 %P 1-24 %G eng %0 Journal Article %J ICML Workshop on Uncertainty in Deep Learning %D 2020 %T Amortised Variational Inference for Hierarchical Mixture Models %A J. Antoran, %A Yao, J. %A Pan, W. %A Doshi-Velez, F. %A J. Hernandez-Lobato %B ICML Workshop on Uncertainty in Deep Learning %P 1-11 %G eng %0 Journal Article %J AAMAS Workshop on Multi-Agent Based Simulation (AAMAS) %D 2020 %T Active Screening on Recurrent Diseases Contact Networks with Uncertainty: a Reinforcement Learning Approach %A HC. Ou, %A Wang, K. %A Doshi-Velez, F. %A M. Tambe %B AAMAS Workshop on Multi-Agent Based Simulation (AAMAS) %P 1-12 %G eng %0 Journal Article %J Advances in Approximate Bayesian Inference %D 2020 %T Characterizing and Avoiding Problematic Global Optima of Variational Autoencoders %A Yacoby, Y. %A Pan, W. %A Doshi-Velez, F. %B Advances in Approximate Bayesian Inference %V 1 %P 1-17 %G eng %0 Journal Article %J ACM Conference on Health, Inference and Learning %D 2020 %T Defining Admissible Rewards for High-Confidence Policy Evaluation in Batch Reinforcement Learning %A Prasad, N. %A Engelhardt, B. %A Doshi-Velez, F. %B ACM Conference on Health, Inference and Learning %V 2 %P 1-9 %G eng %0 Journal Article %J AISTATS %D 2020 %T Prediction Focused Topic Models via Feature Selection %A J. Ren %A Kunes, R. %A Doshi-Velez, F. %B AISTATS %V 2 %P 1-19 %G eng %0 Journal Article %J AISTATS %D 2020 %T POPCORN: Partially Observed Prediction Constrained Reinforcement Learning %A Futoma, J. %A Hughes, M. %A Doshi-Velez, F. %B AISTATS %V 2 %P 1-18 %G eng %0 Journal Article %J AAAI %D 2020 %T Regional Tree Regularization for Interpretability in Deep Neural Networks %A Wu, M. %A S. Parbhoo %A Hughes, M. %A Kindle, R. %A Celi, L. %A Zazzi, M. %A Volker, R. %A Doshi-Velez, F. %B AAAI %V 3 %P 1-9 %G eng %0 Journal Article %J AAAI %D 2020 %T Regional Tree Regularization for Interpretability in Deep Neural Networks %A Wu, M. %A S. Parbhoo %A Hughes, M. %A Kindle, R. %A Celi, L. %A Zazzi, M. %A Volker, R. %A Doshi-Velez, F. %B AAAI %V 3 %P 1-9 %G eng %0 Journal Article %J AAAI %D 2020 %T Ensembles of Locally Independent Prediction Models %A Ross, A. %A Pan, W. %A Celi, L. %A Doshi-Velez, F. %B AAAI %V 3 %P 1-11 %G eng %0 Journal Article %J AMIA CRI %D 2020 %T Identifying Distinct, Effective Treatments for Acute Hypotension with SODA-RL: Safely Optimized Diverse Accurate Reinforcement Learning %A Futoma, F. %A Masgood, M. %A Doshi-Velez, F. %B AMIA CRI %V 1 %P 1-24 %G eng %0 Journal Article %J AMIA CRI %D 2020 %T Interpretable Batch IRL to extract clinician goals in ICU HypotensionManagement %A Srinivasan, S. %A Doshi-Velez, F. %B AMIA CRI %P 636-645 %G eng %0 Journal Article %J Pediatrics %D 2020 %T Big Data in the Assessment of Pediatric Medication Safety %A McMahon, A. %A Cooper, W. %A J. Brown %A Carleton, B. %A Doshi-Velez, F. %A Kohane, I. %A Goldman, J. %A Hoffman, M. %A Kamaleswaran, R. %A Sakiyama, M. %A Sekine, S. %A Sturkenboom, M. %A Turner, M. %A Califf, R. %B Pediatrics %V 145 %P 1-11 %G eng %N 2 %0 Journal Article %J JAMA %D 2020 %T Evaluating Machine Learning Articles %A Doshi-Velez, F. %A Perlis, R. %B JAMA %V 322 %P 1777-1779 %G eng %N 18 %0 Journal Article %J Translational Psychiatry %D 2020 %T Predicting treatment dropout after antidepressant initiation %A Pradier, M. %A McCoy, T. %A Hughes, M. %A Perlis, R. %A Doshi-Velez, F. %B Translational Psychiatry %V 10 %P 1-8 %G eng %N 1 %0 Journal Article %J Nature Medicine %D 2019 %T Do no harm: a roadmap for responsible machine learning for health care %A Wiens, Jenna %A Saria, Suchi %A Sendak, Mark %A Marzyeh Ghassemi %A Liu, Vincent X. %A Finale Doshi-Velez %A Jung, Kenneth %A Heller, Katherine %A Kale, David %A Saeed, Mohammed %A Ossorio, Pilar N. %A Thadaney-Israni, Sonoo %A Goldenberg, Anna %K Health care %K Research management %K Scientific community %X Interest in machine-learning applications within medicine has been growing, but few studies have progressed to deployment in patient care. We present a framework, context and ultimately guidelines for accelerating the translation of machine-learning-based interventions in health care. To be successful, translation will require a team of engaged stakeholders and a systematic process from beginning (problem formulation) to end (widespread deployment). %B Nature Medicine %V 25 %P 1337–1340 %G eng %U https://www.nature.com/articles/s41591-019-0548-6 %N 9 %R 10.1038/s41591-019-0548-6 %0 Conference Paper %B proceedings at The Conference on Uncertainty in Artificial Intelligence (UAI) %D 2019 %T Towards Expressive Priors for Bayesian Neural Networks: Poisson Process Radial Basis Function Networks %A Coker, B. %A Pradier, M. %A Doshi-Velez, F. %B proceedings at The Conference on Uncertainty in Artificial Intelligence (UAI) %V 1 %P 1-37 %G eng %0 Journal Article %J Journal of Machine Learning Research %D 2019 %T A Particle-Based Variational Approach to Bayesian Non-negative Matrix Factorization %A M. Masood %A Doshi-Velez, F. %B Journal of Machine Learning Research %V 20 %P 1-56 %G eng %N 90 %0 Journal Article %J Journal of Machine Learning Research %D 2019 %T Model Selection in Bayesian Neural Networks via Horseshoe Priors %A Ghosh, S. %A Yao, J. %A Doshi-Velez, F. %B Journal of Machine Learning Research %V 20 %P 1-46 %G eng %N 182 %0 Journal Article %J NeurIPS Workshop on Safety and Robustness in Decision-Making, %D 2019 %T Defining Admissible Rewards for High Confidence Policy Evaluation %A Prasad, N. %A Engelhardt, B. %A Doshi-Velez, F. %B NeurIPS Workshop on Safety and Robustness in Decision-Making, %V 1 %P 1-12 %G eng %0 Journal Article %J NeurIPS Workshop on Bayesian Deep Learning %D 2019 %T Controlled Direct Effect Priors for Bayesian Neural Networks %A Ross, A. %A Du, J. %A Sharvit, Y. %A Doshi-Velez, F. %B NeurIPS Workshop on Bayesian Deep Learning %V 1 %P 1-8 %G eng %0 Journal Article %J CSCW Workshop: Identifying Challenges and Opportunities in Human–AI Collaboration in Healthcare %D 2019 %T Integrating AI Recommendations into The Pharmacologic Management of Major Depressive Disorder %A Jacobs, M. %A Perlis, R. %A Pradier, M. %A Doshi-Velez, F. %A Mynatt, E. %A Gajos, K. %B CSCW Workshop: Identifying Challenges and Opportunities in Human–AI Collaboration in Healthcare %V 1 %P 1-5 %G eng %0 Journal Article %J NeurIPS Workshop on Human-Centric ML %D 2019 %T Prediction Focused Topic Models Via Vocab Filtering %A J. Ren %A Russell, R. %A Doshi-Velez, F. %B NeurIPS Workshop on Human-Centric ML %V 1 %P 1-12 %G eng %0 Journal Article %J Advances in Approximate Bayesian Inference %D 2019 %T Challenges in Computing and Optimizing Upper Bounds of Marginal Likelihood based on Chi-Square Divergences %A Pradier, M. %A Hughes, M. %A Doshi-Velez, F. %B Advances in Approximate Bayesian Inference %V 1 %P 1-11 %G eng %0 Journal Article %J ACML Workshop on Weakly Supervised Learning Workshop %D 2019 %T Projected BNNs: Avoiding Weight-space Pathologies by Learning Latent Representations of Neural Network Weights %A Pradier, M. %A Pan, W. %A Yao, J. %A Ghosh, S. %A Doshi-Velez, F. %B ACML Workshop on Weakly Supervised Learning Workshop %V 3 %P 1-15 %G eng %0 Journal Article %J NeurIPS Workshop on Machine Learning for Health %D 2019 %T Prediction Focused Topic Models for Electronic Health Records %A J. Ren %A Kunes, R. %A Doshi-Velez, F. %B NeurIPS Workshop on Machine Learning for Health %V 1 %P 1-13 %G eng %0 Journal Article %J Nature Medicine %D 2019 %T Do no harm: A roadmap for responsible machine learning for healthcare %A Wiens, J. %A Saria, S. %A Sendak, M. %A Ghassemi, M. %A Liu, V. %A Doshi-Velez, F. %A Jung, K. %A Heller, K. %A Kale, D. %A Saeed, M. %A Ossorio, P. %A Thadaney-Israni, S. %A Goldenberg, A. %B Nature Medicine %V 25 %P 1337-1340 %G eng %N 10 %0 Journal Article %J Journal of Autonomous Agents and Multi-Agent Systems (AAMAS) %D 2019 %T Summarizing Agent Strategies %A Amir, O. %A Doshi-Velez, F. %A Sarne, D. %B Journal of Autonomous Agents and Multi-Agent Systems (AAMAS) %V 33 %P 628-644 %G eng %0 Conference Paper %B in proceedings at the American Society for Reproductive Medicine Scientific Congress & Expo (ASRM) %D 2019 %T The Application of Machine Learning Methods to Evaluate Predictors for Live Birth in Programmed Thaw Cycles %A Vaughan, D. %A Pan, W. %A Yacoby, Y. %A Seidler, E. %A Leung, A. %A Doshi-Velez, F. %A Sakkas, D. %B in proceedings at the American Society for Reproductive Medicine Scientific Congress & Expo (ASRM) %G eng %0 Conference Paper %B proceedings at the International Conference on Machine Learning: Workshop on Understanding and Improving Generalization in Deep Learning(ICML) %D 2019 %T Output-Constrained Bayesian Neural Network %A W. Yang %A Lorch, L. %A Graule, M. %A Srinivasan, S. %A Suresh, A. %A Yao, J. %A Pradier, M. %A Doshi-Velez, F. %B proceedings at the International Conference on Machine Learning: Workshop on Understanding and Improving Generalization in Deep Learning(ICML) %G eng %0 Conference Paper %B proceedings at the International Conference on Machine Learning: Workshop on Uncertainty & Robustness in Deep Learning (ICML) %D 2019 %T Output-Constrained Bayesian Neural Networks %A W. Yang %A Lorch, L. %A Graule, M. %A Srinivasan, S. %A Suresh, A. %A Yao, J. %A Pradier, M. %A Doshi-Velez, F. %B proceedings at the International Conference on Machine Learning: Workshop on Uncertainty & Robustness in Deep Learning (ICML) %G eng %0 Conference Paper %B proceedings at the International Conference on Machine Learning: Workshop on Uncertainty & Robustness in Deep Learning (ICML) %D 2019 %T Mitigating Model Non-Identifiability in BNN with Latent Variables %A Yacoby, Y. %A Pan, W. %A Doshi-Velez, F. %B proceedings at the International Conference on Machine Learning: Workshop on Uncertainty & Robustness in Deep Learning (ICML) %G eng %0 Conference Paper %B proceedings at the International Conference on Machine Learning: Workshop on Uncertainty & Robustness in Deep Learning (ICML) %D 2019 %T Quality of Uncertainty Quantification for Bayesian Neural Network Inference %A Yao, J. %A Pan, W. %A Ghosh, S. %A Doshi-Velez, F. %B proceedings at the International Conference on Machine Learning: Workshop on Uncertainty & Robustness in Deep Learning (ICML) %G eng %0 Conference Paper %B proceedings at the International Conference on Bayesian Nonparametrics (BNP) %D 2019 %T Poisson Process Bayesian Neural Networks %A Coker, B. %A Pradier, M. %A Doshi-Velez, F. %B proceedings at the International Conference on Bayesian Nonparametrics (BNP) %G eng %0 Conference Paper %B proceedings at the International Joint Conference on Artificial Intelligence (IJCAI) %D 2019 %T Toward Robust Policy Summarization %A I. Lage %A Lifschitz, D. %A Doshi-Velez, F. %A Amir, O. %B proceedings at the International Joint Conference on Artificial Intelligence (IJCAI) %G eng %0 Conference Paper %B proceedings at the International Conference on Autonomous Agents and Multiagent Systems (AAMAS) %D 2019 %T Toward Robust Summarization of Agent Policies %A I. Lage %A Lifschitz, D. %A Doshi-Velez, F. %A Amir, O. %B proceedings at the International Conference on Autonomous Agents and Multiagent Systems (AAMAS) %G eng %0 Conference Paper %B proceedings at the 7th AAAI Conference on Human Computation and Crowdsourcing (HCOMP) %D 2019 %T Human Evaluation of Models Built for Interpretability %A I. Lage %A E. Chen %A J. He %A M. Narayanan %A B. Kim %A S. Gershman %A Doshi-Velez, F. %B proceedings at the 7th AAAI Conference on Human Computation and Crowdsourcing (HCOMP) %G eng %0 Conference Paper %B proceedings at the International Joint Conference on Artificial Intelligence (IJCAI) %D 2019 %T Truly Batch Apprenticeship Learning with Deep Successor Features %A Srinivasan, S. %A D. Lee %A Doshi-Velez, F. %B proceedings at the International Joint Conference on Artificial Intelligence (IJCAI) %G eng %0 Conference Paper %B proceedings at the International Joint Conference on Artificial Intelligence: Workshop on Explainable Artificial Intelligence (IJCAI), %D 2019 %T Exploring Computational User Models for Agent Policy Summarization %A Lage, I %A Lifschitz, D. %A Doshi-Velez, F. %A Amir, O. %B proceedings at the International Joint Conference on Artificial Intelligence: Workshop on Explainable Artificial Intelligence (IJCAI), %G eng %0 Conference Paper %B in proceedings at the International Joint Conference on Artificial Intelligence. A Workshop on Explainable Artificial Intelligence. %D 2019 %T Explainable Reinforcement Learning via Reward Decomposition %A Z. Juozapaitis %A A. Koul %A A. Fern %A M. Erwig %A Doshi-Velez, F. %B in proceedings at the International Joint Conference on Artificial Intelligence. A Workshop on Explainable Artificial Intelligence. %G eng %0 Conference Paper %B proceedings at the International Joint Conference on Artificial Intelligence (IJCAI) %D 2019 %T Diversity-Inducing Policy Gradient: Using Maximum Mean Discrepancy to Find a Set of Diverse Policies %A M. Masood %A Doshi Velez, F. %B proceedings at the International Joint Conference on Artificial Intelligence (IJCAI) %G eng %0 Conference Paper %B International Conference on Machine Learning (IMCL). %D 2019 %T Combining Parametric and Nonparametric Models for off-policy evaluation %A Gottesman, O. %A Liu, Y. %A Susser, E. %A Brunskill, E. %A Doshi-Velez, F. %B International Conference on Machine Learning (IMCL). %G eng %0 Journal Article %J Statistical Analysis and Data Mining %D 2019 %T Assessing topic model relevance: Evaluation and informative priors %A Angela Fan %A Finale Doshi-Velez %A Miratrix, Luke %B Statistical Analysis and Data Mining %V 12 %P 210-222 %G eng %0 Journal Article %J Nature Medicine %D 2019 %T Guidelines for reinforcement learning in healthcare %A Omer Gottesman %A Johansson, Fredrik %A Matthieu Komorowski %A Aldo Faisal %A Sontag, David %A Finale Doshi-Velez %A Celi, Leo %B Nature Medicine %V 25 %P 16-18 %G eng %0 Conference Proceedings %B Conference on Neural Information Processing Systems (NeurIPS) Workshop on Correcting and Critiquing Trends in Machine Learning %D 2018 %T An Evaluation of the Human-Interpretability of Explanation %A I. Lage %A E. Chen %A J. He %A M. Narayanan %A S. Gershman %A B. Kim %A Doshi-Velez, F. %B Conference on Neural Information Processing Systems (NeurIPS) Workshop on Correcting and Critiquing Trends in Machine Learning %G eng %0 Conference Proceedings %B Conference on Neural Information Processing Systems (NeurIPS) Workshop on Bayesian Deep Learning %D 2018 %T Projected BNNs: Avoiding weight-space pathologies by projecting neural network weights %A Melanie F. Pradier %A Weiwei Pan %A Jiayu Yao %A Soumya Ghosh %A Finale Doshi-Velez %B Conference on Neural Information Processing Systems (NeurIPS) Workshop on Bayesian Deep Learning %G eng %0 Conference Proceedings %B Conference on Neural Information Processing Systems (NeurIPS) Workshop on All of Bayesian Nonparametrics %D 2018 %T Hierarchical Stick-breaking Feature Paintbox %A M. Fernandez-Pradier %A Pan, W. %A M. Yao %A R. Singh %A Finale Doshi-Velez %B Conference on Neural Information Processing Systems (NeurIPS) Workshop on All of Bayesian Nonparametrics %G eng %0 Conference Proceedings %B Conference on Neural Information Processing Systems (NeurIPS) Workshop on Reinforcement Learning under Partial Observability %D 2018 %T Prediction-Constrained POMDPs %A Joseph Futoma %A Michael C. Hughes %A Finale Doshi-Velez %B Conference on Neural Information Processing Systems (NeurIPS) Workshop on Reinforcement Learning under Partial Observability %G eng %0 Journal Article %J PLoS ONE %D 2018 %T Improving counterfactual reasoning with kernelised dynamic mixing models %A Parbhoo, Sonali %A Omer Gottesman %A Andrew Slavin Ross %A Matthieu Komorowski %A Aldo Faisal %A Bon, Isabella %A Roth, Volker %A Finale Doshi-Velez %B PLoS ONE %V 13 %G eng %N 11 %0 Conference Proceedings %B Conference on Neural Information Processing Systems (NeurIPS) %D 2018 %T Human-in-the-Loop Interpretability Prior %A Isaac Lage %A Andrew Ross %A Kim, Been %A Gershman, Samuel %A Finale Doshi-Velez %B Conference on Neural Information Processing Systems (NeurIPS) %G eng %0 Conference Proceedings %B Association for the Advancement of Artificial Intelligence (AAAI) %D 2018 %T Beyond Sparsity: Tree Regularization of Deep Models for Interpretability %A Mike Wu %A Michael Hughes %A Parbhoo, Sonali %A Zazzi, Maurizio %A Roth, Volker %A Finale Doshi-Velez %B Association for the Advancement of Artificial Intelligence (AAAI) %G eng %0 Conference Proceedings %B International Conference on Machine Learning (ICML) Exploration in Reinforcement Learning Workshop %D 2018 %T Diversity-Inducing Policy Gradient: Using MMD to find a set of policies that are diverse in terms of stete-visitation %A Muhammad A Masood %A Finale Doshi-Velez %B International Conference on Machine Learning (ICML) Exploration in Reinforcement Learning Workshop %G eng %0 Conference Proceedings %B American Medical Informatics Association (AMIA) Annual Symposium %D 2018 %T Improving Sepsis Treatment Strategies by Combining Deep and Kernel-Based Reinforcement Learning %A Xuefeng Peng %A Yi Ding %A David Wihl %A Omer Gottesman %A Matthieu Komorowski %A Li-wei H. Lehman %A Andrew Ross %A Aldo Faisal %A Finale Doshi-Velez %B American Medical Informatics Association (AMIA) Annual Symposium %G eng %0 Conference Paper %B Proceedings of the 35th International Conference on Machine Learning (ICML) %D 2018 %T Structured Variational Learning of Bayesian Neural Networks with Horseshoe Priors %A Soumya Ghosh %A Jiayu Yao %A Finale Doshi-Velez %B Proceedings of the 35th International Conference on Machine Learning (ICML) %C Stockholm, Sweden %V 80 %G eng %0 Conference Paper %B Proceedings of the 35th International Conference on Machine Learning (ICML) %D 2018 %T Decomposition of Uncertainty in Bayesian Deep Learning for Efficient and Risk-sensitive Learning %A Depeweg, Stefan %A Jose Miguel Hernandez-Lobato %A Finale Doshi-Velez %A Steffen Udluft %B Proceedings of the 35th International Conference on Machine Learning (ICML) %C Stockholm, Sweden %V 80 %G eng %0 Conference Paper %B Proceedings of the 21st International Conference on Artificial Intelligence and Statistics (AISTATS) 2018 %D 2018 %T Weighted Tensor Decomposition for Learning Latent Variables with Partial Data %A Omer Gottesman %A Weiewei Pan %A Finale Doshi-Velez %B Proceedings of the 21st International Conference on Artificial Intelligence and Statistics (AISTATS) 2018 %C Lanzarote, Spain %V 84 %G eng %0 Conference Paper %B Proceedings of the 21st International Conference on Artificial Intelligence and Statistics (AISTATS) 2018 %D 2018 %T Semi-Supervised Prediction-Constrained Topic Models %A Michael C. Hughes %A Gabriel Hope %A Leah Weiner %A Thomas H. McCoy, Jr. %A Roy H. Perlis %A Erik Sudderth %A Finale Doshi-Velez %B Proceedings of the 21st International Conference on Artificial Intelligence and Statistics (AISTATS) 2018 %C Lanzarote, Spain %V 84 %G eng %0 Conference Proceedings %B Autonomous Agents and Multiagent Systems, Blue Sky Ideas Track %D 2018 %T Agent Strategy Summarization %A Amir, Ofra %A Finale Doshi-Velez %A David Sarne %B Autonomous Agents and Multiagent Systems, Blue Sky Ideas Track %G eng %0 Conference Proceedings %B Privacy Law Scholars Conference %D 2018 %T Accountability of AI Under the Law: The Role of Explanation %A Finale Doshi-Velez %A Mason Kortz %A Ryan Budish %A Chris Bavitz %A Sam Gershman %A David O'Brien %A Shieber, Stuart %A James Waldo %A David Weinberger %A Alexandra Wood %B Privacy Law Scholars Conference %G eng %0 Conference Proceedings %B Association for Computational Linguistics %D 2018 %T Unsupervised Grammar Induction with Depth-bounded PCFG %A Lifeng Jin %A Finale Doshi-Velez %A Miller, Timothy %A Schuler, William %A Schwartz, Lane %B Association for Computational Linguistics %G eng %0 Conference Proceedings %B Association for the Advancement of Artificial Intelligence (AAAI) %D 2018 %T Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing their Input Gradients %A Andrew Slavin Ross %A Finale Doshi-Velez %B Association for the Advancement of Artificial Intelligence (AAAI) %G eng %0 Conference Proceedings %B International Conference on Machine Learning (ICML) Workshop on Lifelong Learning, %D 2018 %T Direct Policy Transfer via Hidden Parameter Markov Decision Processes %A Jiayu Yao %A Taylor Killian %A George Konidaris %A Finale Doshi-Velez %B International Conference on Machine Learning (ICML) Workshop on Lifelong Learning, %G eng %0 Conference Proceedings %B International Conference on Machine Learning (ICML) Workshop on CausalML %D 2018 %T Behaviour Policy Estimation in Off-Policy Policy Evaluation: Calibration Matters %A Aniruddh Raghu %A Omer Gottesman %A Yao Liu %A Matthieu Komorowski %A Aldo Faisal %A Finale Doshi-Velez %A Emma Brunskill %B International Conference on Machine Learning (ICML) Workshop on CausalML %G eng %0 Conference Proceedings %B International Conference on Machine Learning (ICML) Workshop on CausalML, %D 2018 %T Stitched Trajectories for Off-Policy Learning %A Scott Sussex %A Omer Gottesman %A Yao Liu %A Susan Murphy %A Emma Brunskill %A Finale Doshi-Velez %B International Conference on Machine Learning (ICML) Workshop on CausalML, %G eng %0 Conference Proceedings %B International Conference on Machine Learning (ICML) Workshop on CausalML %D 2018 %T Representation Balancing MDPs for Off-Policy Policy Evaluation %A Yao Liu %A Omer Gottesman %A Aniruddh Raghu %A Matthieu Komorowski %A Aldo Faisal %A Finale Doshi-Velez %A Emma Brunskill %B International Conference on Machine Learning (ICML) Workshop on CausalML %G eng %0 Conference Proceedings %B International Conference on Machine Learning (ICML) Exploration in Reinforcement Learning Workshop, %D 2018 %T Regularizing Tensor Decomposition Methods by Optimizing Pseudo-Data %A Omer Gottesman %A Finale Doshi-Velez %B International Conference on Machine Learning (ICML) Exploration in Reinforcement Learning Workshop, %G eng %0 Conference Proceedings %B International Conference on Machine Learning (ICML) Workshop on Human Interpretability in Machine Learning, %D 2018 %T Learning Qualitatively Diverse and Interpretable Rules for Classification %A Andrew Slavin Ross %A Weiwei Pan %A Finale Doshi-Velez %B International Conference on Machine Learning (ICML) Workshop on Human Interpretability in Machine Learning, %G eng %0 Conference Proceedings %B Conference on Empirical Methods in Natural Language Processing (EMNLP) %D 2018 %T Depth-bounding is effective: Improvements and Evaluation of Unsupervised PCFG Induction %A Lifeng Jin %A Finale Doshi-Velez %A Miller, Tim %A Schuler, William %A Schwartz, Lane %B Conference on Empirical Methods in Natural Language Processing (EMNLP) %G eng %0 Book Section %B Explainable and Interpretable Models in Computer Vision and Machine Learning %D 2018 %T Considerations for Evaluation and Generalization in Interpretable Machine Learning %A Finale Doshi-Velez %A Kim, Been %E Hugo Escalante %E Sergio Escalera %E Isabelle Guyon %E Xavier Baró %E Yağmur Güçlütürk %E Umut Güçlü %E Marcel A. J. van Gerven %B Explainable and Interpretable Models in Computer Vision and Machine Learning %7 1st %I Springer International Publishing %G eng %0 Journal Article %J IEEE Transactions on Visualization and Computer Graphics %D 2018 %T PhenoLines: Phenotype Comparison Visualizations for Disease Subtyping via Topic Models %A Michael Glueck %A Mahdi Pakdaman Naeini %A Finale Doshi-Velez %A Fanny Chevalier %A Azam Khan %A Daniel Wigdor %A Brudno, Michael %B IEEE Transactions on Visualization and Computer Graphics %V 24 %P 371-381 %G eng %N 1 %0 Conference Proceedings %B American Medical Informatics Association (AMIA), %D 2017 %T Predicting intervention onset in the ICU with switching state space models %A Marzyeh Ghassemi %A Mike Wu %A Michael C. Hughes %A Szolovits, Peter %A Finale Doshi-Velez %B American Medical Informatics Association (AMIA), %G eng %0 Conference Proceedings %B Conference on Visual Analytics Science and Technology (VAST), %D 2017 %T PhenoLines: Phenotype Comparison Visualizations for Disease Subtyping via Topic Models %A Michael Glueck %A Mahdi Pakdaman Naeini %A Finale Doshi-Velez %A Fanny Chevalier %A Azam Khan %A Daniel Wigdor %A Brudno, Michael %B Conference on Visual Analytics Science and Technology (VAST), %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) Workshop on Transparent and Interpretable Machine Learning in Safety Critical Environments %D 2017 %T The Neural LASSO: Local Linear Sparsity for Interpretable Explanations %A Andrew Slavin Ross %A Isaac Lage %A Finale Doshi-Velez %B Neural Information Processing Systems (NIPS) Workshop on Transparent and Interpretable Machine Learning in Safety Critical Environments %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) Workshop on Transparent and Interpretable Machine Learning in Safety Critical Environments %D 2017 %T Beyond Sparsity: Tree Regularization of Deep Models for Interpretability %A Mike Wu %A Michael Hughes %A Parbhoo, Sonali %A Zazzi, Maurizio %A Roth, Volker %A Finale Doshi-Velez %B Neural Information Processing Systems (NIPS) Workshop on Transparent and Interpretable Machine Learning in Safety Critical Environments %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) Workshop on Machine Learning for Healthcare %D 2017 %T Counterfactual Reasoning with Dynamic Switching Models for HIV Therapy Selection %A S. Parbhoo %A V. Roth %A Finale Doshi-Velez %B Neural Information Processing Systems (NIPS) Workshop on Machine Learning for Healthcare %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) Workshop on Machine Learning for Healthcare %D 2017 %T Prediction-Constrained Topic Models for Antidepressant Recommendation %A Michael C. Hughes %A Gabriel Hope %A Leah Weiner %A Thomas H. McCoy %A Roy H. Perlis %A Erik B. Sudderth %A Finale Doshi-Velez %B Neural Information Processing Systems (NIPS) Workshop on Machine Learning for Healthcare %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) Workshop on Bayesian Deep Learning %D 2017 %T Model Selection in Bayesian Neural Networks via Horseshoe Priors %A Soumya Ghosh %A Finale Doshi-Velez %B Neural Information Processing Systems (NIPS) Workshop on Bayesian Deep Learning %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) Workshop on Advances in Approximate Bayesian Inference %D 2017 %T Structured Variational Autoencoders for the Beta-Bernoulli Process %A Rachit Singh %A Jeffrey Ling %A Finale Doshi-Velez %B Neural Information Processing Systems (NIPS) Workshop on Advances in Approximate Bayesian Inference %G eng %0 Conference Proceedings %B Advancing Education with Data Knowledge Discovery and Data Mining (KDD) Workshop %D 2017 %T Clustering LaTeX Solutions to Machine Learning Assignments for Rapid Assessment %A Sindy Tan %A Finale Doshi-Velez %A Juan Quiroz %A Elena Glassman %B Advancing Education with Data Knowledge Discovery and Data Mining (KDD) Workshop %G eng %0 Conference Proceedings %B International Conference on Machine Learning (ICML) Workshop %D 2017 %T Uncertainty Decomposition in Bayesian Neural Networks with Latent Variables %A Depeweg, Stefan %A Jose Miguel Hernandez-Lobato %A Finale Doshi-Velez %A Steffen Udluft %B International Conference on Machine Learning (ICML) Workshop %G eng %0 Conference Paper %B AMIA Summits on Translational Science Proceedings %D 2017 %T Combining Kernel and Model Based Learning for HIV Therapy Selection %A Parbhoo, Sonali %A Bogojeska, Jasmina %A Zazzi, Maurizio %A Roth, Volker %A Finale Doshi-Velez %B AMIA Summits on Translational Science Proceedings %V 2017 %P 239 %G eng %0 Conference Proceedings %B Text as Data %D 2017 %T Prior Matters: Simple and General Methods for Evaluating and Improving Topic Quality in Topic Modeling %A Angela Fan %A Finale Doshi-Velez %A Miratrix, Luke %B Text as Data %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) %D 2017 %T Robust and Efficient Transfer Learning with Hidden Parameter Markov Decision Processes %A Taylor Killian %A Samuel Daulton %A George Konidaris %A Finale Doshi-Velez %B Neural Information Processing Systems (NIPS) %C Long Beach, CA %G eng %0 Journal Article %J Journal of Machine Learning %D 2017 %T A Bayesian Framework for Learning Rule Sets for Interpretable Classification %A Wang, Tong %A Rudin, Cynthia %A Finale Doshi-Velez %A Liu, Yimin %A Klampfl, Erica %A MacNeille, Perry %B Journal of Machine Learning %V 18 %P 1-37 %G eng %N 70 %0 Conference Proceedings %B International Joint Conference on Artificial Intelligence (IJCAI) %D 2017 %T Right for the Right Reasons: Training Differentiable Models by Constraining their Explananations %A Andrew Slavin Ross %A Michael C. Hughes %A Finale Doshi-Velez %B International Joint Conference on Artificial Intelligence (IJCAI) %C Melbourne, Australia %G eng %0 Journal Article %J Statistics and Computing %D 2017 %T Restricted Indian Buffet Processes %A Finale Doshi-Velez %A Sinead Williamson %B Statistics and Computing %V 27 %P 1205-1223 %G eng %N 5 %0 Journal Article %J Journal of the American Medical Informatics Association %D 2017 %T Understanding Vasopressor Intervention and Weaning: Risk Prediction in a Public Heterogeneous Clinical Time Series Database %A Mike Wu %A Marzyeh Ghassemi %A Mengling Fend %A Leo A. Celi %A Szolovits, Peter %A Finale Doshi-Velez %X

 

Background The widespread adoption of electronic health records allows us to ask evidence-based questions about the need for and benefits of specific clinical interventions in critical-care settings across large populations.

Objective We investigated the prediction of vasopressor administration and weaning in the intensive care unit. Vasopressors are commonly used to control hypotension, and changes in timing and dosage can have a large impact on patient outcomes.

Materials and Methods We considered a cohort of 15 695 intensive care unit patients without orders for reduced care who were alive 30 days post-discharge. A switching-state autoregressive model (SSAM) was trained to predict the multidimensional physiological time series of patients before, during, and after vasopressor administration. The latent states from the SSAM were used as predictors of vasopressor administration and weaning.

Results The unsupervised SSAM features were able to predict patient vasopressor administration and successful patient weaning. Features derived from the SSAM achieved areas under the receiver operating curve of 0.92, 0.88, and 0.71 for predicting ungapped vasopressor administration, gapped vasopressor administration, and vasopressor weaning, respectively. We also demonstrated many cases where our model predicted weaning well in advance of a successful wean.

Conclusion Models that used SSAM features increased performance on both predictive tasks. These improvements may reflect an underlying, and ultimately predictive, latent state detectable from the physiological time series.

 

%B Journal of the American Medical Informatics Association %V 24 %P 488-495 %G eng %N 3 %0 Journal Article %J ICLR %D 2017 %T Learning and Policy Search in Stochastic Dynamical Systems with Bayesian Neural Networks %A Stefan Depewag %A José Miguel Hernández-Lobato %A Finale Doshi-Velez %A Steffen Udluft %X

We present an algorithm for model-based reinforcement learning that combines Bayesian neural networks (BNNs) with random roll-outs and stochastic optimization for policy learning. The BNNs are trained by minimizing α  -divergences, allowing us to capture complicated statistical patterns in the transition dynamics, e.g. multi-modality and heteroskedasticity, which are usually missed by other common modeling approaches. We illustrate the performance of our method by solving a challenging benchmark where model-based approaches usually fail and by obtaining promising results in a real-world scenario for controlling a gas turbine.

%B ICLR %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) Workshop for Machine Learning and Healthcare %D 2016 %T Combining Kernel and Model Based Learning for HIV Therapy Selection %A Parbhoo, Sonali %A Bogojeska, Jasmina %A Zazzi, Maurizio %A Roth, Volker %A Finale Doshi-Velez %B Neural Information Processing Systems (NIPS) Workshop for Machine Learning and Healthcare %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) Workshop for Machine Learning and Healthcare %D 2016 %T Transfer Learning Across Patient Variations with Hidden Parameter Markov Decision Processes %A Taylor W. Killian %A George Konidaris %A Finale Doshi-Velez %B Neural Information Processing Systems (NIPS) Workshop for Machine Learning and Healthcare %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) Workshop for Machine Learning and Healthcare %D 2016 %T Supervised topic models for clinical interpretability %A Michael C. Hughes %A Elibol, Huseyin Melih %A Thomas McCoy %A Roy Perlis %A Finale Doshi-Velez %B Neural Information Processing Systems (NIPS) Workshop for Machine Learning and Healthcare %G eng %0 Conference Proceedings %B International Conference on Machine Learning (ICML) Workshop on Geometry in Machine Learning %D 2016 %T Robust Posterior Exploration in NMF %A Muhammad Arjumand Masood %A Finale Doshi-Velez %B International Conference on Machine Learning (ICML) Workshop on Geometry in Machine Learning %G eng %0 Conference Proceedings %B Computational Linguistics: Technical Papers (COLING) %D 2016 %T Memory-Bounded Left-Corner Unsupervised Grammar Induction on Child-Directed Input %A Cory Shain %A William Bryce %A Lifeng Jin %A Victoria Krakovna %A Finale Doshi-Velez %A Miller, Timothy %A Schuler, William %A Schwartz, Lane %B Computational Linguistics: Technical Papers (COLING) %I COLING %C Osaka, Japan %P 964-975 %G eng %0 Conference Proceedings %B IJCAI %D 2016 %T Hidden Parameter Markov Decision Processes: A Semiparametric Regression Approach for Discovering Latent Task Parametrizations %A Doshi-Velez, F. %A Konidaris, G %B IJCAI %G eng %0 Journal Article %J Journal of Machine Learning Research %D 2016 %T Cross-Corpora Unsupervised Learning of Trajectories in Autism Spectrum Disorders %A Melih Elibol %A Vincent Nguyen %A Scott Linderman %A Matthew Johnson %A Amna Hashmi %A Finale Doshi-Velez %B Journal of Machine Learning Research %V 17 %P 4597-4634 %G eng %N 1 %0 Journal Article %J AISTATS %D 2016 %T Spectral M-estimation with Application to Hidden Markov Models: Supplementary Material %A Dustin Tran %A Minjae Kim %A Finale Doshi-Velz %B AISTATS %G eng %0 Journal Article %J arXiv:1604.00653 %D 2016 %T A Characterization of the Non-Uniqueness of Nonnegative Matrix Factorizations %A Weiwei Pan %A Finale Doshi-Velez %X

Nonnegative matrix factorization (NMF) is a popular dimension reduction technique that produces interpretable decomposition of the data into parts. However, this decompostion is not generally identifiable (even up to permutation and scaling). While other studies have provide criteria under which NMF is identifiable, we present the first (to our knowledge) characterization of the non-identifiability of NMF. We describe exactly when and how non-uniqueness can occur, which has important implications for algorithms to efficiently discover alternate solutions, if they exist.

%B arXiv:1604.00653 %G eng %0 Journal Article %D 2016 %T Cost-Sensitive Batch Mode Active Learning: Designing Astronomical Observation by Optimizing Telescope Time and Telescope Choice %A Xide Xia %A Pavlos Protopapas %A Finale Doshi-Velez %X

%G eng %0 Journal Article %J arXiv preprint arXiv:1606.06250 %D 2016 %T An Empirical Comparison of Sampling Quality Metrics: A Case Study for Bayesian Nonnegative Matrix Factorization %A Arjumand Masood %A Weiwei Pan %A Finale Doshi-Velez %X

In this work, we empirically explore the question: how can we assess the quality of samples from some target distribution? We assume that the samples are provided by some valid Monte Carlo procedure, so we are guaranteed that the collection of samples will asymptotically approximate the true distribution. Most current evaluation approaches focus on two questions: (1) Has the chain mixed, that is, is it sampling from the distribution? and (2) How independent are the samples (as MCMC procedures produce correlated samples)? Focusing on the case of Bayesian nonnegative matrix factorization, we empirically evaluate standard metrics of sampler quality as well as propose new metrics to capture aspects that these measures fail to expose. The aspect of sampling that is of particular interest to us is the ability (or inability) of sampling methods to move between multiple optima in NMF problems. As a proxy, we propose and study a number of metrics that might quantify the diversity of a set of NMF factorizations obtained by a sampler through quantifying the coverage of the posterior distribution. We compare the performance of a number of standard sampling methods for NMF in terms of these new metrics.

%B arXiv preprint arXiv:1606.06250 %G eng %0 Journal Article %J Sensors and Actuators A: Physical %D 2016 %T Machine Learning Approaches to Environmental Disturbance Rejection in Multi-Axis Optoelectronic Force Sensors %A Joshua Gafford %A Finale Doshi-Velez %A Robert Wood %A Walsh, Conor %X

Light-intensity modulated (LIM) force sensors are seeing increasing interest in the field of surgical robotics and flexible systems in particular. However, such sensing modalities are notoriously susceptible to ambient effects such as temperature and environmental irradiance which can register as false force readings. We explore machine learning techniques to dynamically compensate for environmental biases that plague multi-axis optoelectronic force sensors. In this work, we fabricate a multisensor: three-axis LIM force sensor with integrated temperature and ambient irradiance sensing manufactured via a monolithic, origami-inspired fabrication process called printed-circuit MEMS. We explore machine learning regression techniques to compensate for temperature and ambient light sensitivity using on-board environmental sensor data. We compare batch-based ridge regression, kernelized regression and support vector techniques to baseline ordinary least-squares estimates to show that on-board environmental monitoring can substantially improve sensor force tracking performance and output stability under variable lighting and large (>100 °C) thermal gradients. By augmenting the least-squares estimate with nonlinear functions describing both environmental disturbances and cross-axis coupling effects, we can reduce the error in Fx, Fy and Fz by 10%, 33%, and 73%, respectively. We assess viability of each algorithm tested in terms of both prediction accuracy and computational overhead, and analyze kernel-based regression for prediction in the context of online force feedback and haptics applications in surgical robotics. Finally, we suggest future work for fast approximation and prediction using stochastic, sparse kernel techniques.

%B Sensors and Actuators A: Physical %V 248 %P 78-87 %G eng %0 Journal Article %J PLoS ONE 11(7): e0159621 %D 2016 %T Electronic Health Record Based Algorithm to Identify Patients with Autism Spectrum Disorder %A Todd Lingren %A Pei Chen %A Joseph Bochenek %A Finale Doshi-Velez %A Patty Manning-Courtney %A Julie Bickel %A Leah Wildenger Welchons %A Judy Reinhold %A Nicole Bing %A Yizhao Ni %A William Barbaresi %A Frank Mentch %A Melissa Basford %A Joshua Denny %A Lyam Vazquez %A Cassandra Perry %A Bahram Namjou %A Haijun Qiu %A John Connolly %A Debra Abrams %A Ingrid A. Holm %A Beth A. Cobb %A Nataline Lingren %A Imre Solti %A Hakonarson, Hakon %A Isaac S. Kohane %A Harley, John %A Savova, Guergana %B PLoS ONE 11(7): e0159621 %G eng %0 Journal Article %J arXiv:1606.05320 %D 2016 %T Increasing the Interpretability of Recurrent Neural Networks Using Hidden Markov Models %A Viktoriya Krakovna %A Finale Doshi-Velez %X

Abstract: As deep neural networks continue to revolutionize various application domains, there is increasing interest in making these powerful models more understandable and interpretable, and narrowing down the causes of good and bad predictions. We focus on recurrent neural networks (RNNs), state of the art models in speech recognition and translation. Our approach to increasing interpretability is by combining an RNN with a hidden Markov model (HMM), a simpler and more transparent model. We explore various combinations of RNNs and HMMs: an HMM trained on LSTM states; a hybrid model where an HMM is trained first, then a small LSTM is given HMM state distributions and trained to fill in gaps in the HMM's performance; and a jointly trained hybrid model. We find that the LSTM and HMM learn complementary information about the features in the text.

%B arXiv:1606.05320 %G eng %0 Journal Article %J arXiv:1504.07614 %D 2015 %T Bayesian Or's of And's for Interpretable Classification with Application to Context Aware Recommender Systems %A Wang, Tong %A Rudin, Cynthia %A Finale Doshi-Velez %A Liu, Yimin %A Klampfl, Erica %A MacNeille, Perry %B arXiv:1504.07614 %G eng %0 Conference Paper %B Advances in Neural Information Processing Systems %D 2015 %T Mind the Gap: A Generative Approach to Interpretable Feature Selection and Extraction %A Kim, Been %A Shah, Julie A %A Finale Doshi-Velez %B Advances in Neural Information Processing Systems %P 2251–2259 %G eng %0 Journal Article %J Inflammatory bowel diseases %D 2015 %T Prevalence of Inflammatory Bowel Disease Among Patients with Autism Spectrum Disorders %A Finale Doshi-Velez %A Avillach, Paul %A Palmer, Nathan %A Bousvaros, Athos %A Yaorong Ge %A Fox, Kathe %A Steinberg, Greg %A Spettell, Claire %A Juster, Iver %A Isaac Kohane %B Inflammatory bowel diseases %I LWW %V 21 %P 2281–2288 %G eng %0 Journal Article %J IEEE Transactions on Pattern Analysis and Machine Intelligence %D 2015 %T Bayesian Nonparametric Methods for Partially-Observable Reinforcement Learning %A Finale Doshi-Velez %A David Pfau %A Frank Wood %A Nicholas Roy %X

Making intelligent decisions from incomplete information is critical in many applications: for example, robots must choose actions based on imperfect sensors, and speech-based interfaces must infer a user’s needs from noisy microphone inputs. What makes these tasks hard is that often we do not have a natural representation with which to model the domain and use for choosing actions; we must learn about the domain’s properties while simultaneously performing the task. Learning a representation also involves trade-offs between modeling the data that we have seen previously and being able to make predictions about new data. This article explores learning representations of stochastic systems using Bayesian nonparametric statistics. Bayesian nonparametric methods allow the sophistication of a representation to scale gracefully with the complexity in the data. Our main contribution is a careful empirical evaluation of how representations learned using Bayesian nonparametric methods compare to other standard learning approaches, especially in support of planning and control. We show that the Bayesian aspects of the methods result in achieving state-of-the-art performance in decision making with relatively few samples, while the nonparametric aspects often result in fewer computations. These results hold across a variety of different techniques for choosing actions given a representation.

%B IEEE Transactions on Pattern Analysis and Machine Intelligence %V 37 %P 394 - 407 %G eng %N 2 %0 Journal Article %D 2015 %T HackEbola with Data: On the hackathon format for timely data analysis. %A Finale Doshi-Velez %A Yael E. Marshall %X For more information, see the event page: https://projects.iq.harvard.edu/hack %G eng %0 Journal Article %J AAAI %D 2015 %T Graph-Sparse LDA: A Topic Model with Structured Sparsity %A Finale Doshi-Velez %A Byron C. Wallace %A Adams, Ryan P %X

Originally designed to model text, topic modeling has become a powerful tool for uncovering latent structure in domains including medicine, finance, and vision. The goals for the model vary depending on the application: in some cases, the discovered topics may be used for prediction or some other downstream task. In other cases, the content of the topic itself may be of intrinsic scientific interest. Unfortunately, even using modern sparse techniques, the discovered topics are often difficult to interpret due to the high dimensionality of the underlying space. To improve topic interpretability, we introduce Graph-Sparse LDA, a hierarchical topic model that leverages knowledge of relationships between words (e.g., as encoded by an ontology). In our model, topics are summarized by a few latent concept-words from the underlying graph that explain the observed words. Graph-Sparse LDA recovers sparse, interpretable summaries on two real-world biomedical datasets while matching state-of-the-art prediction performance.

%B AAAI %G eng %0 Journal Article %J Pediatrics %D 2014 %T Comorbidity clusters in autism spectrum disorders: an electronic health record time-series analysis %A Finale Doshi-Velez %A Yaorong Ge %A Isaac Kohane %B Pediatrics %I American Academy of Pediatrics %V 133 %P e54–e63 %G eng %0 Journal Article %J arXiv:1410.4510 %D 2014 %T Graph-Sparse LDA: A Topic Model with Structured Sparsity %A Finale Doshi-Velez %A Wallace, Byron %A Adams, Ryan %B arXiv:1410.4510 %G eng %0 Conference Proceedings %B AAAI 2014 Fall Symposium on Knowledge, Skill, and Behavior Transfer in Autonomous Robots %D 2014 %T Hidden Parameter Markov Decision Processes: An Emerging Paradigm for Modeling Families of Related Tasks %A George Konidaris %A Finale Doshi-Velez %X

The goal of transfer is to use knowledge obtained by solving one task to improve a robot’s (or software agent’s) performance in future tasks. In general, we do not expect this to work; for transfer to be feasible, there must be something in common between the source task(s) and goal task(s). The question at the core of the transfer learning enterprise is therefore: what makes two tasks related?, or more generally, how do you define a family of related tasks? Given a precise definition of how a particular family of tasks is related, we can formulate clear optimization methods for selecting source tasks and determining what knowledge should be imported from the source task(s), and how it should be used in the target task(s). This paper describes one model that has appeared in several different research scenarios where an agent is faced with a family of tasks that have similar, but not identical, dynamics (or reward functions). For example, a human learning to play baseball may, over the course of their career, be exposed to several different bats, each with slightly different weights and lengths. A human who has learned to play baseball well with one bat would be expected to be able to pick up any similar bat and use it. Similarly, when learning to drive a car, one may learn in more than one car, and then be expected to be able to drive any make and model of car (within reasonable variations) with little or no relearning. These examples are instances of exactly the kind of flexible, reliable, and sample-efficient behavior that we should be aiming to achieve in robotics applications. One way to model such a family of tasks is to posit that they are generated by a small set of latent parameters (e.g., the length and weight of the bat, or parameters describing the various physical properties of the car’s steering system and clutch) that are fixed for each problem instance (e.g., for each bat, or car), but are not directly observable by the agent. Defining a distribution over these latent parameters results in a family of related tasks, and transfer is feasible to the extent that the number of latent variables is small, the task dynamics (or reward function) vary smoothly with them, and to the extent to which they can either be ignored or identified using transition data from the task. This model has appeared under several different names in the literature; we refer to it as a hidden-parameter Markov decision process (or HIPMDP).

%B AAAI 2014 Fall Symposium on Knowledge, Skill, and Behavior Transfer in Autonomous Robots %G eng %0 Conference Proceedings %B ACM SIGKDD international conference on Knowledge discovery and data mining %D 2014 %T Unfolding Physiological State: Mortality Modelling in Intensive Care Units %A Marzyeh Ghassemi %A Tristan Naumann %A Finale Doshi-Velez %A Nicole Brimmer %A Rohit Joshi %A Anna Rumshisky %A Szolovits, Peter %X

Accurate knowledge of a patient’s disease state and trajectory is critical in a clinical setting. Modern electronic healthcare records contain an increasingly large amount of data, and the ability to automatically identify the factors that influence patient outcomes stand to greatly improve the ef- ficiency and quality of care. We examined the use of latent variable models (viz. Latent Dirichlet Allocation) to decompose free-text hospital notes into meaningful features, and the predictive power of these features for patient mortality. We considered three prediction regimes: (1) baseline prediction, (2) dynamic (timevarying) outcome prediction, and (3) retrospective outcome prediction. In each, our prediction task differs from the familiar time-varying situation whereby data accumulates; since fewer patients have long ICU stays, as we move forward in time fewer patients are available and the prediction task becomes increasingly difficult. We found that latent topic-derived features were effective in determining patient mortality under three timelines: inhospital, 30 day post-discharge, and 1 year post-discharge mortality. Our results demonstrated that the latent topic features important in predicting hospital mortality are very different from those that are important in post-discharge. mortality. In general, latent topic features were more predictive than structured features, and a combination of the two performed best. The time-varying models that combined latent topic features and baseline features had AUCs that reached 0.85, 0.80, and 0.77 for in-hospital, 30 day post-discharge and 1 year post-discharge mortality respectively. Our results agreed with other work suggesting that the first 24 hours of patient information are often the most predictive of hospital mortality. Retrospective models that used a combination of latent topic features and structured features achieved AUCs of 0.96, 0.82, and 0.81 for in-hospital, 30 day, and 1-year mortality prediction. Our work focuses on the dynamic (time-varying) setting, because models from this regime could facilitate an on-going severity stratification system that helps d

%B ACM SIGKDD international conference on Knowledge discovery and data mining %C New York City %P 75-84 %G eng %0 Journal Article %J Pediatrics %D 2013 %T Comorbidity Clusters in Autism Spectrum Disorders: An Electronic Health Record Time-Series Analysis %A Finale Doshi-Velez %A Yaorong Ge %A Isaac Kohane %X

OBJECTIVE: The distinct trajectories of patients with autism spectrum disorders (ASDs) have not been extensively studied, particularly regarding clinical manifestations beyond the neurobehavioral criteria from the Diagnostic and Statistical Manual of Mental Disorders. The objective of this study was to investigate the patterns of co-occurrence of medical comorbidities in ASDs.

METHODS: International Classification of Diseases, Ninth Revision codes from patients aged at least 15 years and a diagnosis of ASD were obtained from electronic medical records. These codes were aggregated by using phenotype-wide association studies categories and processed into 1350-dimensional vectors describing the counts of the most common categories in 6-month blocks between the ages of 0 to 15. Hierarchical clustering was used to identify subgroups with distinct courses.

RESULTS: Four subgroups were identified. The first was characterized by seizures (n = 120, subgroup prevalence 77.5%). The second (n = 197) was characterized by multisystem disorders including gastrointestinal disorders (prevalence 24.3%) and auditory disorders and infections (prevalence 87.8%), and the third was characterized by psychiatric disorders (n = 212, prevalence 33.0%). The last group (n = 4316) could not be further resolved. The prevalence of psychiatric disorders was uncorrelated with seizure activity (P = .17), but a significant correlation existed between gastrointestinal disorders and seizures (P < .001). The correlation results were replicated by using a second sample of 496 individuals from a different geographic region.

CONCLUSIONS: Three distinct patterns of medical trajectories were identified by unsupervised clustering of electronic health record diagnoses. These may point to distinct etiologies with different genetic and environmental contributions. Additional clinical and molecular characterizations will be required to further delineate these subgroups.

%B Pediatrics %V 10.1542 %P 0819 %G eng %N peds.2013 %0 Journal Article %J CoRR %D 2013 %T Hidden Parameter Markov Decision Processes: A Semiparametric Regression Approach for Discovering Latent Task Parametrizations %A Finale Doshi-Velez %A George Konidaris %X

Control applications often feature tasks with similar, but not identical, dynamics. We introduce the Hidden Parameter Markov Decision Process (HiP-MDP), a framework that parametrizes a family of related dynamical systems with a low-dimensional set of latent factors, and introduce a semiparametric regression approach for learning its structure from data. In the control setting, we show that a learned HiP-MDP rapidly identifies the dynamics of a new task instance, allowing an agent to flexibly adapt to task variations.

%B CoRR %V abs/1308.3513 %G eng %0 Thesis %D 2012 %T Bayesian nonparametric approaches for reinforcement learning in partially observable domains %A Finale Doshi-Velez %I Massachusetts Institute of Technology %G eng %9 phd %0 Journal Article %J Journal of the American Medical Directors Association %D 2012 %T Improving safety and operational efficiency in residential care settings with WiFi-based localization %A Finale Doshi-Velez %A William Li %A Yoni Battat %A Ben Charrow %A Dorothy Curthis %A Jun-Geun Park %A Hemachandra, Sachithra %A Javier Velez %A Cynthia Walsh %A Don Fredette %A others %B Journal of the American Medical Directors Association %I Elsevier %V 13 %P 558–563 %G eng %0 Journal Article %J the NIPS 2012 Workshop on Bayesian Nonparametric Models for Reliable Planning And Decision-Making Under Uncertainty %D 2012 %T Transfer Learning by Discovering Latent Task Parametrizations %A Finale Doshi-Velez %A George Konidaris %B the NIPS 2012 Workshop on Bayesian Nonparametric Models for Reliable Planning And Decision-Making Under Uncertainty %G eng %0 Journal Article %J IEEE International Conference on Robotics and Automation %D 2012 %T A Bayesian nonparametric approach to modeling battery health %A Joshua Mason Joseph %A Finale Doshi-Velez %A Nicholas Roy %X

Abstract—Making intelligent decisions from incomplete information is critical in many applications: for example, robots must choose actions based on imperfect sensors, and speech-based interfaces must infer a user’s needs from noisy microphone inputs. What makes these tasks hard is that often we do not have a natural representation with which to model the domain and use for choosing actions; we must learn about the domain’s properties while simultaneously performing the task. Learning a representation also involves trade-offs between modeling the data that we have seen previously and being able to make predictions about new data. This article explores learning representations of stochastic systems using Bayesian nonparametric statistics. Bayesian nonparametric methods allow the sophistication of a representation to scale gracefully with the complexity in the data. Our main contribution is a careful empirical evaluation of how representations learned using Bayesian nonparametric methods compare to other standard learning approaches, especially in support of planning and control. We show that the Bayesian aspects of the methods result in achieving state-of-the-art performance in decision making with relatively few samples, while the nonparametric aspects often result in fewer computations. These results hold across a variety of different techniques for choosing actions given a representation. Index Terms—Artificial intelligence, machine learning, reinforcement learning, partially-observable Markov decision process, hierarchial Dirichlet process hidden Markov model.

%B IEEE International Conference on Robotics and Automation %P 1876–1882 %G eng %R 10.1109/ICRA.2012.6225178 %0 Conference Proceedings %B 33rd Annual Meeting of the Cognitive Science Society (CogSci) %D 2011 %T A Comparison of Human and Agent Reinforcement Learning in Partially Observable Domains %A Finale Doshi-Velez %A Zoubin Ghahramani %B 33rd Annual Meeting of the Cognitive Science Society (CogSci) %G eng %0 Conference Proceedings %B Neural Information Processing Systems (NIPS) Workshop: Bayesian Nonparametrics, Hope or Hype? %D 2011 %T An Analysis of Activity Changes in MS Patients: A Case Study in the Use of Bayesian Nonparametrics %A Finale Doshi-Velez %A Nicholas Roy %B Neural Information Processing Systems (NIPS) Workshop: Bayesian Nonparametrics, Hope or Hype? %G eng %0 Journal Article %J Auton. Robots %D 2011 %T A Bayesian nonparametric approach to modeling motion patterns %A Joshua Mason Joseph %A Finale Doshi-Velez %A Albert S. Huang %A Nicholas Roy %B Auton. Robots %V 31 %P 383–400 %G eng %R 10.1007/s10514-011-9248-x %0 Conference Paper %B Proceedings of the 28th International Conference on Machine Learning %D 2011 %T Infinite Dynamic Bayesian Networks %A Finale Doshi %A David Wingate %A Joshua B. Tenenbaum %A Nicholas Roy %B Proceedings of the 28th International Conference on Machine Learning %P 913–920 %G eng %0 Conference Paper %B Proceedings of the 28th International Conference on Machine Learning %D 2011 %T Online Discovery of Feature Dependencies %A Alborz Geramifard %A Finale Doshi %A Josh Redding %A Nicholas Roy %A Jonathan P. How %B Proceedings of the 28th International Conference on Machine Learning %P 881–888 %G eng %0 Conference Paper %B Proceedings of the Twenty-Fourth Conference on Artificial Intelligence %D 2010 %T A Bayesian Nonparametric Approach to Modeling Mobility Patterns %A Joshua Mason Joseph %A Finale Doshi-Velez %A Nicholas Roy %B Proceedings of the Twenty-Fourth Conference on Artificial Intelligence %G eng %0 Conference Paper %B Conference on Artificial Intelligence %D 2010 %T Nonparametric Bayesian Approaches for Reinforcement Learning in Partially Observable Domains %A Finale Doshi-Velez %B Conference on Artificial Intelligence %G eng %0 Conference Paper %B Advances in Neural Information Processing Systems 23: 24th Annual Conference on Neural Information Processing Systems 2010. Proceedings of a meeting held 6-9 December 2010, Vancouver, British Columbia, Canada. %D 2010 %T Nonparametric Bayesian Policy Priors for Reinforcement Learning %A Finale Doshi-Velez %A David Wingate %A Nicholas Roy %A Joshua B. Tenenbaum %B Advances in Neural Information Processing Systems 23: 24th Annual Conference on Neural Information Processing Systems 2010. Proceedings of a meeting held 6-9 December 2010, Vancouver, British Columbia, Canada. %P 532–540 %G eng %0 Conference Proceedings %B Advances in Neural Information Processing Systems (NIPS) %D 2009 %T The Infinite Partially Observable Markov Decision Process %A Finale Doshi-Velez %B Advances in Neural Information Processing Systems (NIPS) %G eng %0 Conference Paper %B Proceedings of the 26th International Conference on Machine Learning %D 2009 %T Accelerated Sampling for the Indian Buffet Process %A Finale Doshi-Velez %A Zoubin Ghahramani %X

We often seek to identify co-occurring hidden features in a set of observations. The Indian Buffet Process (IBP) provides a nonparametric prior on the features present in each observation, but current inference techniques for the IBP often scale poorly. The collapsed Gibbs sampler for the IBP has a running time cubic in the number of observations, and the uncollapsed Gibbs sampler, while linear, is often slow to mix. We present a new linear-time collapsed Gibbs sampler for conjugate likelihood models and demonstrate its efficacy on large real-world datasets.

%B Proceedings of the 26th International Conference on Machine Learning %C Montreal, Canada %G eng %0 Conference Paper %B {UAI} 2009, Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence, Montreal, QC, Canada, June 18-21, 2009 %D 2009 %T Correlated Non-Parametric Latent Feature Models %A Finale Doshi-Velez %A Zoubin Ghahramani %B {UAI} 2009, Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence, Montreal, QC, Canada, June 18-21, 2009 %P 143–150 %G eng %0 Conference Proceedings %B Conference on Neural Information Processing Systems (NIPS) %D 2009 %T Large Scale Nonparametric Bayesian Inference: Data Parallelisation in the Indian Buffet Process %A Finale Doshi-Velez %A David A. Knowles %A Shakir Mohamed %A Zoubin Ghahramani %B Conference on Neural Information Processing Systems (NIPS) %G eng %0 Conference Proceedings %B Artificial Intelligence on Statistics (AISTATS) Best Paper Nominee %D 2009 %T Variational Inference for the Indian Buffet Process %A Finale Doshi %A Kurt Miller %A Jurgen Van Gael %A Yee Whye Teh %B Artificial Intelligence on Statistics (AISTATS) Best Paper Nominee %G eng %0 Journal Article %J Connection Science %D 2008 %T Spoken Language Interaction with Model Uncertainty: An Adaptive Human-Robot Interaction System %A Finale Doshia %A Nicholas Roy %X

Spoken language is one of the most intuitive forms of interaction between humans and agents. Unfortunately, agents that interact with people using natural language often experience communication errors and do not correctly understand the user’s intentions. Recent systems have successfully used probabilistic models of speech, language, and user behavior to generate robust dialog performance in the presence of noisy speech recognition and ambiguous language choices, but decisions made using these probabilistic models are still prone to errors due to the complexity of acquiring and maintaining a complete model of human language and behavior. In this paper, we describe a decision-theoretic model for human-robot interaction using natural language. Our algorithm is based on the Partially Observable Markov Decision Process (POMDP), which allows agents to choose actions that are robust not only to uncertainty from noisy or ambiguous speech recognition but also unknown user models. Like most dialog systems, a POMDP is defined by a large number of parameters that may be difficult to specify a priori from domain knowledge, and learning these parameters from the user may require an unacceptably long training period. We describe an extension to the POMDP model that allows the agent to acquire a linguistic model of the user online, including new vocabulary and word choice preferences. Our approach not only avoids a training period of constant questioning as the agent learns, but also allows the agent to actively query for additional information when its uncertainty suggests a high risk of mistakes. We demonstrate our approach both in simulation and on a natural language interaction system for a robotic wheelchair application. Keywords: dialog management, human-computer interface, adaptive systems, online learning, partially observable Markov decision processes

%B Connection Science %V 20 %P 299-318 %G eng %N 4 %0 Conference Proceedings %B Conference on Autonomous Agents and Multi-Agent Systems (AAMAS) Best Paper Nominee %D 2008 %T The Permutable POMDP: Fast Solutions to POMDPs for Preference Elicitation %A Finale Doshi %A Nicholas Roy %B Conference on Autonomous Agents and Multi-Agent Systems (AAMAS) Best Paper Nominee %G eng %R 10.1145/1402383.1402454 %0 Conference Proceedings %B International Conference on Machine Learning (ICML) %D 2008 %T Reinforcement Learning with Limited Reinforcement: Using Bayes Risk for Active Learning in POMDPs %A Finale Doshi %A Joelle Pineau %A Nicholas Roy %B International Conference on Machine Learning (ICML) %G eng %R 10.1145/1390156.1390189 %0 Conference Paper %B Symposium on Language and Robotics (LANGRO) %D 2007 %T Learning User Models with Limited Reinforcement: An Adaptive Human-Robot Interaction System %A Finale Doshi %A Nicholas Roy %B Symposium on Language and Robotics (LANGRO) %G eng %0 Conference Proceedings %B Conference on Human Robot Interaction (HRI) %D 2007 %T Efficient Model Learning for Dialog Management %A Finale Doshi %A Nicholas Roy %B Conference on Human Robot Interaction (HRI) %G eng %0 Conference Proceedings %B Conference on Intelligent Robots and Systems (IROS) %D 2007 %T Collision Detection in Legged Locomotion Using Supervised Learning %A Finale Doshi %A Emma Brunskill %A Alexander C. Shkolnik %A Thomas Kollar %A Khashayar Rohanimanesh %A Russ Tedrake %A Nicholas Roy %B Conference on Intelligent Robots and Systems (IROS) %C San Diego CA %G eng %R 10.1109/IROS.2007.4399538