Sebastian Thrun’s research while affiliated with Stanford University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (520)


MAPTree: Beating “Optimal” Decision Trees with Bayesian Decision Trees
  • Article

March 2024

·

17 Reads

Proceedings of the AAAI Conference on Artificial Intelligence

·

Mo Tiwari

·

Sebastian Thrun

Decision trees remain one of the most popular machine learning models today, largely due to their out-of-the-box performance and interpretability. In this work, we present a Bayesian approach to decision tree induction via maximum a posteriori inference of a posterior distribution over trees. We first demonstrate a connection between maximum a posteriori inference of decision trees and AND/OR search. Using this connection, we propose an AND/OR search algorithm, dubbed MAPTree, which is able to recover the maximum a posteriori tree. Lastly, we demonstrate the empirical performance of the maximum a posteriori tree both on synthetic data and in real world settings. On 16 real world datasets, MAPTree either outperforms baselines or demonstrates comparable performance but with much smaller trees. On a synthetic dataset, MAPTree also demonstrates greater robustness to noise and better generalization than existing approaches. Finally, MAPTree recovers the maxiumum a posteriori tree faster than existing sampling approaches and, in contrast with those algorithms, is able to provide a certificate of optimality. The code for our experiments is available at https://github.com/ThrunGroup/maptree.



Bayesian Decision Trees via Tractable Priors and Probabilistic Context-Free Grammars

February 2023

·

20 Reads

Decision Trees are some of the most popular machine learning models today due to their out-of-the-box performance and interpretability. Often, Decision Trees models are constructed greedily in a top-down fashion via heuristic search criteria, such as Gini impurity or entropy. However, trees constructed in this manner are sensitive to minor fluctuations in training data and are prone to overfitting. In contrast, Bayesian approaches to tree construction formulate the selection process as a posterior inference problem; such approaches are more stable and provide greater theoretical guarantees. However, generating Bayesian Decision Trees usually requires sampling from complex, multimodal posterior distributions. Current Markov Chain Monte Carlo-based approaches for sampling Bayesian Decision Trees are prone to mode collapse and long mixing times, which makes them impractical. In this paper, we propose a new criterion for training Bayesian Decision Trees. Our criterion gives rise to BCART-PCFG, which can efficiently sample decision trees from a posterior distribution across trees given the data and find the maximum a posteriori (MAP) tree. Learning the posterior and training the sampler can be done in time that is polynomial in the dataset size. Once the posterior has been learned, trees can be sampled efficiently (linearly in the number of nodes). At the core of our method is a reduction of sampling the posterior to sampling a derivation from a probabilistic context-free grammar. We find that trees sampled via BCART-PCFG perform comparable to or better than greedily-constructed Decision Trees in classification accuracy on several datasets. Additionally, the trees sampled via BCART-PCFG are significantly smaller -- sometimes by as much as 20x.


MABSplit: Faster Forest Training Using Multi-Armed Bandits

December 2022

·

23 Reads

Random forests are some of the most widely used machine learning models today, especially in domains that necessitate interpretability. We present an algorithm that accelerates the training of random forests and other popular tree-based learning methods. At the core of our algorithm is a novel node-splitting subroutine, dubbed MABSplit, used to efficiently find split points when constructing decision trees. Our algorithm borrows techniques from the multi-armed bandit literature to judiciously determine how to allocate samples and computational power across candidate split points. We provide theoretical guarantees that MABSplit improves the sample complexity of each node split from linear to logarithmic in the number of data points. In some settings, MABSplit leads to 100x faster training (an 99% reduction in training time) without any decrease in generalization performance. We demonstrate similar speedups when MABSplit is used across a variety of forest-based variants, such as Extremely Random Forests and Random Patches. We also show our algorithm can be used in both classification and regression tasks. Finally, we show that MABSplit outperforms existing methods in generalization performance and feature importance calculations under a fixed computational budget. All of our experimental results are reproducible via a one-line script at https://github.com/ThrunGroup/FastForest.


Differentiation of Active Corneal Infections From Healed Scars Using Deep Learning

August 2021

·

71 Reads

·

28 Citations

Ophthalmology

Objective To develop and evaluate an automated, portable algorithm to differentiate active corneal ulcers from healed scars using only external photographs. Design A convolutional neural network was trained and tested using photographs of corneal ulcers and scars. Subjects De-identified photographs of corneal ulcers were obtained from the Steroids for Corneal Ulcers Trial (SCUT), Mycotic Ulcer Treatment Trial (MUTT), and Byers Eye Institute at Stanford University. Methods Photographs of corneal ulcers (n=1313) and scars (n=1132) from the SCUT and MUTT trials were used to train a convolutional neural network (CNN). The CNN was tested on two different patient populations from eye clinics in India (n=200) and the Byers Eye Institute at Stanford University (n=101). Accuracy was evaluated against gold standard clinical classifications. Feature importances for the trained model were visualized using Gradient-weighted Class Activation Mapping (Grad-CAM). Main Outcome Measure Accuracy of the CNN was assessed via F1 score. Area under the receiver operating characteristic curve (ROC) was used to measure the precision-recall trade-off. Results The CNN correctly classified 115/123 active ulcers and 65/77 scars in corneal ulcer patients from India (F1 score: 92.0% (95% CI: 88.2 - 95.8%), sensitivity: 93.5% (95% CI: 89.1 - 97.9%), specificity: 84.42% (95% CI: 79.42 - 89.42%), ROC (AUC=0.9731)). The CNN correctly classified 43/55 active ulcers and 42/46 scars in corneal ulcer patients from Northern California (F1 score: 84.3% (95% CI: 77.2 - 91.4%), sensitivity: 78.2% (95% CI: 67.3 - 89.1%), specificity: 91.3% (95% CI: 85.8 - 96.8%), ROC (AUC=0.9474)). The CNN visualizations correlated with clinically relevant features such as corneal infiltrate, hypopyon, and conjunctival injection. Conclusion The CNN classified corneal ulcers and scars with high accuracy and generalizes to patient populations outside of its training data. The CNN focuses on clinically relevant features when it makes a diagnosis. The CNN demonstrates potential as an inexpensive diagnostic approach that may aid triage in communities with limited access to eye care.


Bandit-PAM: Almost Linear Time k-Medoids Clustering via Multi-Armed Bandits

June 2020

·

82 Reads

Clustering is a ubiquitous task in data science. Compared to the commonly used k-means clustering algorithm, k-medoids clustering algorithms require the cluster centers to be actual data points and support arbitrary distance metrics, allowing for greater interpretability and the clustering of structured objects. Current state-of-the-art k-medoids clustering algorithms, such as Partitioning Around Medoids (PAM), are iterative and are quadratic in the dataset size n for each iteration, being prohibitively expensive for large datasets. We propose Bandit-PAM, a randomized algorithm inspired by techniques from multi-armed bandits, that significantly improves the computational efficiency of PAM. We theoretically prove that Bandit-PAM reduces the complexity of each PAM iteration from O(n2)O(n^2) to O(nlogn)O(n \log n) and returns the same results with high probability, under assumptions on the data that often hold in practice. We empirically validate our results on several large-scale real-world datasets, including a coding exercise submissions dataset from Code.org, the 10x Genomics 68k PBMC single-cell RNA sequencing dataset, and the MNIST handwritten digits dataset. We observe that Bandit-PAM returns the same results as PAM while performing up to 200x fewer distance computations. The improvements demonstrated by Bandit-PAM enable k-medoids clustering on a wide range of applications, including identifying cell types in large-scale single-cell data and providing scalable feedback for students learning computer science online. We also release Python and C++ implementations of our algorithm.


Artificial intelligence and cancer

February 2020

·

160 Reads

·

35 Citations

Nature Cancer

Filtered through the analytical power of artificial intelligence, the wealth of available biomedical data promises to revolutionize cancer research, diagnosis and care. In this Viewpoint, six experts discuss some of the challenges, exciting developments and future questions arising at the interface of machine learning and oncology.


Fig. 1 | Deep learning. a, A simple, multilayer deep neural network takes two classes of data, denoted by the different colors, and makes them linearly separable by iteratively distorting the data as it flows from layer to layer. The final output layer serves as a classifier by outputting the probability of either one of the classes. This example illustrates the basic concept used by large scale networks. Conceptual illustration adapted with permission from http://colah.github.io/. b, Example large-scale network that accepts as input a variety of data types (images, time-series, etc.), and for each data type learns a useful featurization in its lowerlevel towers. The data from each tower is then merged and flows through higher levels, allowing the DNN to perform inference across data types-a capability that is increasingly important in healthcare.
A guide to deep learning in healthcare
  • Article
  • Full-text available

January 2019

·

29,210 Reads

·

3,022 Citations

Nature Medicine

Here we present deep-learning techniques for healthcare, centering our discussion on deep learning in computer vision, natural language processing, reinforcement learning, and generalized methods. We describe how these computational techniques can impact a few key areas of medicine and explore how to build end-to-end systems. Our discussion of computer vision focuses largely on medical imaging, and we describe the application of natural language processing to domains such as electronic health record data. Similarly, reinforcement learning is discussed in the context of robotic-assisted surgery, and generalized deep-learning methods for genomics are reviewed.

Download

Corrigendum: Dermatologist-level classification of skin cancer with deep neural networks

June 2017

·

338 Reads

·

260 Citations

Nature

In the Acknowledgements section of this Letter, the sentence: "This study was supported by the Baxter Foundation, California Institute for Regenerative Medicine (CIRM) grants TT3-05501 and RB5-07469 and US National Institutes of Health (NIH) grants AG044815, AG009521, NS089533, AR063963 and AG020961 (H.M.B.)" should have read: "This study was supported by funding from the Baxter Foundation to H.M.B." Furthermore, the last line of the Acknowledgements section should have read: "In addition, this work was supported by a National Institutes of Health (NIH) National Center for Advancing Translational Science Clinical and Translational Science Award (UL1 TR001085). The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIH." The original Letter has been corrected online. © 2017 Macmillan Publishers Limited, part of Springer Nature. All rights reserved.


Dermatologist-level classification of skin cancer with deep neural networks

January 2017

·

5,895 Reads

·

11,828 Citations

Nature

Skin cancer, the most common human malignancy, is primarily diagnosed visually, beginning with an initial clinical screening and followed potentially by dermoscopic analysis, a biopsy and histopathological examination. Automated classification of skin lesions using images is a challenging task owing to the fine-grained variability in the appearance of skin lesions. Deep convolutional neural networks (CNNs) show potential for general and highly variable tasks across many fine-grained object categories. Here we demonstrate classification of skin lesions using a single CNN, trained end-to-end from images directly, using only pixels and disease labels as inputs. We train a CNN using a dataset of 129,450 clinical images-two orders of magnitude larger than previous datasets-consisting of 2,032 different diseases. We test its performance against 21 board-certified dermatologists on biopsy-proven clinical images with two critical binary classification use cases: keratinocyte carcinomas versus benign seborrheic keratoses; and malignant melanomas versus benign nevi. The first case represents the identification of the most common cancers, the second represents the identification of the deadliest skin cancer. The CNN achieves performance on par with all tested experts across both tasks, demonstrating an artificial intelligence capable of classifying skin cancer with a level of competence comparable to dermatologists. Outfitted with deep neural networks, mobile devices can potentially extend the reach of dermatologists outside of the clinic. It is projected that 6.3 billion smartphone subscriptions will exist by the year 2021 (ref. 13) and can therefore potentially provide low-cost universal access to vital diagnostic care.


Citations (89)


... 3D reconstruction has seen significant advancements recently [Habermann et al. 2019[Habermann et al. , 2020Jiang et al. 2022;Li et al. 2021;Loper et al. 2015;Luo et al. 2023Luo et al. , 2021Poole et al. 2022;Xu et al. 2018;Yan et al. 2024]. Some single-view human reconstruction methods [Anguelov et al. 2005;Hasler et al. 2009;Lassner et al. 2017;Pavlakos et al. 2019;] restrict the solution space to a parametric human model and simplify the problem, which can only reconstruct nude human 3D models without garments. Inspired by SMPL [Loper et al. 2015], some methods [Alldieck et al. 2019;Tan et al. 2020;Xiang et al. 2020;Yang et al. 2018;Zheng et al. 2019] approximate human body geometry by deforming the SMPL. ...

Reference:

GarVerseLOD: High-Fidelity 3D Garment Reconstruction from a Single In-the-Wild Image using a Dataset with Levels of Details
SCAPE: Shape Completion and Animation of People
  • Citing Chapter
  • August 2023

... Parametric models operate under the assumption that the object contour is specified by a pre-fixed number of parameters. For instance, common models in this class assume that the object contour is either the perimeter of a rectangle [13][14][15] or an ellipse, i.e. two curves fully characterized by a position vector, an orientation angle and two semi-axis lengths (totaling five parameters). A more general model that encompasses both is the superellipse [16][17][18], which describes the perimeter of the contour curve using a third parameter to control convexity. ...

Model Based Vehicle Tracking for Autonomous Driving in Urban Environments
  • Citing Chapter
  • June 2009

... The first scheme involves generating dense point clouds based on dense image matching with large overlap, and then using the iterative closest point (ICP) algorithm [17] or its improved versions [18,19] to match the dense point cloud with the original point cloud. Essentially, this scheme converts the 2D-3D data fusion problem into a point cloud registration problem, which is widely used for registration between aerial photographs and point clouds [20][21][22]. ...

Generalized-ICP
  • Citing Chapter
  • July 2010

... Бактеріальний кератитодна з основних причин погіршення або втрати зору в усьому світі. Характеризується швидким перебігом захворювання і досить часто агресивною мікрофлорою, що потребує негайного призначення антибіотиків широкого спектру дії та лікування в умовах офтальмологічного стаціонару.Однак зростання резистентності до антибактеріальних препаратів продовжує збільшуватись, що робить багато широко використовуваних терапевтичних засобів неефективними.Науковці докладають значних зусиль, щоб зрозуміти основний патогенез звичайних мікроорганізмів, які спричиняють розвиток кератиту та умови, які сприяють формуванню антибіотикорезистентності. Все це стимулює удосконалення діагностичних методів,розробку нових методів лікування, покращення клінічних результатів та зменшення інвалідизуючих ускладнень [1,2,3,4]. ...

Differentiation of Active Corneal Infections From Healed Scars Using Deep Learning
  • Citing Article
  • August 2021

Ophthalmology

... Sin embargo, los sistemas producen importantes estrategias terapéuticas más efectivas según una diversidad de imágenes de biopsias más ordenadas. De hecho, hay máquinas que estudian las imágenes sin tener que teñir las células de un paciente y aun así obtienen suficiente información que les sirve como biomarcadores, especialmente útil en un entorno con escasos recursos [15]. ...

Artificial intelligence and cancer
  • Citing Article
  • February 2020

Nature Cancer

... This method explicitly captures energy-based dependencies in humanrobot interactions, enhancing the modeling of cooperative distributed control in dynamic environments. Socially Aware Robot Navigation: Advancements in robotics and artificial intelligence [50] have led to the development of numerous socially-aware robot navigation planners [11]- [13], [20], evolving from early systems like MINERVA [51] and RHINO [52]. These planners can be categorized into decoupled and coupled strategies for HRI inference and path planning [53]. ...

The Museum Tour-Guide Robot RHINO
  • Citing Book
  • January 2011

... DL employs multilayer neural networks that learn data representations through iterative mathematical operations. It excels in processing large-scale data efficiently, demonstrating successes across diverse fields like medical imaging [15], natural language processing [16], sentiment analysis [17], healthcare [18], and image recognition [19]. Convolutional Neural Networks (CNNs) play a vital role in DL, particularly in performing feature extraction [20]. ...

A guide to deep learning in healthcare

Nature Medicine

... Recent indoor MMS use SLAM for indoor localization and mapping, which originates from the robotics and computer vision community (Stachniss et al., 2016;Thrun, 2002). Lehtola et al. (2017) provide a comparison of numerous state-of-the-art LiDAR-focused indoor MMS based on point clouds. ...

Springer Handbook of Robotics, 2nd edition
  • Citing Chapter
  • January 2016

... Meanwhile, consistency regularization may struggle to maintain performance in the presence of strong noise or outliers, particularly in scenarios where the data volume is limited, raising concerns about model stability [13]. Furthermore, existing methods often face difficulties in effectively capturing information across different spatial scales in unlabeled medical images, further limiting their practical applicability [14]. ...

Corrigendum: Dermatologist-level classification of skin cancer with deep neural networks
  • Citing Article
  • June 2017

Nature