PreprintPDF Available

DeepRod: A human-in-the-loop system for automatic rodent behavior analysis

Authors:
Preprints and early-stage research may not have been peer reviewed yet.

Abstract and Figures

We present a human-in-the-loop system for efficient rodent behavior analysis in drug development. Addressing the time-consuming and labor-intensive nature of manual behavior categorization, this UX-optimized platform integrates AI for complex behavior prediction and active learning to identify rare events. The proposed solution leverages a cloud-native data processing pipeline, AI-based novelty behavior recognition and multi-class classification, demonstrating significant improvements in behavior labeling and discovery.
Content may be subject to copyright.
DeepRod: A human-in-the-loop system for automatic rodent behavior analysis
A. Loy1, M. Garafolj1, H. Schauerte2*, H. Behnke1, C. Charnier2, P. Schwarz3, G. Rast2and T. Wollmann1*
1Merantix Momentum GmbH
2Boehringer Ingelheim GmbH & Co. KG, Drug Discovery Sciences
3Boehringer Ingelheim GmbH & Co. KG, BI X
*thomas.wollmann@merantix.com,heike.schauerte@boehringer-ingelheim.com (corresponding authors)
Abstract: We present a human-in-the-loop system for efficient rodent behavior analysis in drug
development. Addressing the time-consuming and labor-intensive nature of manual behavior
categorization, this UX-optimized platform integrates AI for complex behavior prediction and active
learning to identify rare events. The proposed solution leverages a cloud-native data processing pipeline,
AI-based novelty behavior recognition and multi-class classification, demonstrating significant
improvements in behavior labeling and discovery.
Introduction
In drug discovery and development, systematic assessment of drug safety using highly regulated preclinical
studies prior to first-in-human clinical trials are mandatory [1] to ensure safety for volunteers and patients. These
assessments allow detailed views on risk, benefit and the therapeutic index of potential future therapeutics. The
evaluations include, among others, standardized functional behavioral studies in rodents [2, 3, 4]. During the
research phase, automated video-based systems (e.g., PhenoTyper) to assess continuous quantitative and
qualitative motor behavior during the active phase of the rodents. Infrared video cameras located in the top unit
of each observation arena populated with one rodent per arena record 14 h of video material per animal. During
each study, groups of animals exposed to an active ingredient at various doses or receiving a placebo are
recorded from the top, generating large video datasets. Events of interest can be very rare and require in depth
analysis of the footage. Manual analysis is not feasible in an appropriate time and with acceptable effort.
Automated analysis of distance moved and animal velocity (e.g. Ethovision XT) provides very sensitive
measures for central nervous effects and general tolerability. However, these features are not discriminative
enough to detect complex events.
In this work, we propose an UX-optimized platform for behavior labeling and analysis integrated into the
workflow, AI-based complex behavior prediction, active learning to find rare events and propose candidates for
new behavioral categories.
Methods
We propose a novel system for analyzing rodent behavior at scale, which combines a user-centered interface with
AI-based behavior prediction, novel behavior recognition and active learning. The system should support rodent
behavioral analysis by automatizing the behavior classification of rodents. To achieve this, the system needs to
enable users in annotating the rodent behavior in video snippets to gather training data, support users in detecting
novel, previously unseen behaviors and finally to automatically classify the behavior of rodents.
System overview
Each routine study performed to profile a research substance includes 35 million frames from typically 28
individual 14h long video clips with 25 fps, recorded by a PhenoTyper camera [5, 6]. To tackle this amount of
data, an efficient, parallelized, and cloud-native data processing pipeline processes these raw video files [7].
First, the pipeline re-encodes the input video for efficiency and storage and registers the video’s metadata into
.CC-BY-NC-ND 4.0 International licensemade available under a
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is
The copyright holder for this preprintthis version posted January 4, 2024. ; https://doi.org/10.1101/2024.01.04.572506doi: bioRxiv preprint
the system. Second, visual information is extracted using a deep learning approach [8] and meaningful features
are further generated using that information. Lastly, a classifier predicts the rodent behavior for each frame based
on those features.
Behavior recognition
The core component of our system is the behavior recognition component. This component uses a two-staged
machine learning pipeline to classify the behavior of rodents. The first stage extracts visual information from the
video stream by localizing nine anatomical landmarks (“keypoints” in the following) of the rodent. Similar to
MARS [9], keypoints correspond to the nose, ears, body center, hips, tail base, tail center and tail end. Our
keypoint extraction method is based on DeepLabCut [8].
The second stage is a classifier that uses features based on the keypoints. We identified that features capturing
the position, pose, and movement are discriminative for characterizing the rodent’s behavior. Per each feature
category, we engineered a range of features based on the keypoints capturing the rodent’s position, pose and
movement (Figure 1). Some features are aggregated within sliding windows of various sizes. We frame the
behavior detection problem as a multi-class classification problem, containing all known behaviors and an extra
class representing any unknown behavior, which is explicitly labeled to not conform to any of the known
behavior. Our method is leveraging XGBoost [12], which is recognized as a good classifier under skewed data
and noise [13]. The tree-based model also enables interpretability like computing feature importance, which is
favorable in life sciences [14]. The system trains new models automatically based on user request. Users receive
a report after training that offers an intuitive overview of model improvements.
Detection and labeling of rare behavior
To enable automatic rodent behavior classification, collection of annotated data is necessary to train the machine-
learning-based classifier. As there are thousands of hours of video material that can be used to create the
annotated training data set, the choice of which video sections to annotate is not trivial and is subjected to time
constraints of the labeling force.
For efficient use of human labeling resources, the system implements a labeling assistant shown in Figure 2. The
labeling assistant leverages an active learning method based on Meal [15] to suggest areas to label across the
whole video material. The active learning methodology can be formulated as an SQL query, which selects a
fixed amount kof currently unlabeled model predictions from the database. We refer to the result of this query
for a given video as the “labeling queue”. The active learning query consists of multiple subqueries, which each
Figure 1.1. Features based on the
distances between keypoints are
relevant to detect interactions.
Figure 1.2. Features based on [10] use
the relative keypoint positions are
relevant to identify typical postures.
Figure 1.3. Features based on [11] use
the keypoint movements to identify
temporal patterns like directed or
undirected motion.
.CC-BY-NC-ND 4.0 International licensemade available under a
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is
The copyright holder for this preprintthis version posted January 4, 2024. ; https://doi.org/10.1101/2024.01.04.572506doi: bioRxiv preprint
have a capped contribution to the overall labeling queue. The subqueries select samples with the following
properties: Label is likely to belong to an underrepresented class in the annotated data set, has high prediction
uncertainty, has high novelty score, is selected randomly.
Figure 2. Main view of annotation view with the labeling assistant. Labelers are presented with the video footage and the
timeline of model-predicted behaviors as well as annotations that are already set. Annotations can be created with a single
click to allow for an efficient process.
Each item in the labeling queue corresponds to a window of interest, which users are expected to annotate. The
order of the queue is randomized to prevent bias of which sections in the video get labeled. Once the queue is
exhausted, labelers are directed to a new video with a fresh queue and therefore iteratively cycle through all
videos. This setup prevents annotations concentrated in only a small subset of videos and ensures that each video
acquires labels for the top kmost relevant sections.
Novel behavior recognition
The rodents in the experiments might demonstrate unusual or novel behaviors due to the effect of the compounds
that they are exposed to. Therefore, the system needs to support users in observing unusual behavior to enable
them to possibly categorize it as a new behavior class. We refer to this problem as novel behavior recognition”
and formulate it as an outlier detection problem. Each frame gets embedded into a low-dimensional feature space
using principal component analysis (PCA) to reduce the dimensionality and the correlation between features.
Then, mean and variance across all labeled samples for each class (i.e. each defined and annotated behavior
type) are computed. With that, we define the novelty score of a frame as the Mahalanobi’s distance [16] to the
closest known distribution.
.CC-BY-NC-ND 4.0 International licensemade available under a
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is
The copyright holder for this preprintthis version posted January 4, 2024. ; https://doi.org/10.1101/2024.01.04.572506doi: bioRxiv preprint
Figure 3. Visualization of the lower dimension feature embedding space used for novelty scoring of a sample. We visualize
the center and variance of the estimated class distributions. Samples are scored according to their distance to the distributions
of labeled samples. Note that the figure contains a subset of randomly sampled 3000 unlabelled behavior points.
Results
The system was evaluated in a pilot with three distinct expert annotators who created 13.862 new annotations. In
this context, one annotation refers to an identified behavior type with a start and end frame. These annotations
are distributed across 226 individual rodents from 16 distinct experiments.
Active learning results
Figure 4. Recall of our system for rare behavior types with different subsets of the dataset. The percentage of coverage of a
novel behavior when annotating a certain percentage of the overall dataset is shown.
.CC-BY-NC-ND 4.0 International licensemade available under a
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is
The copyright holder for this preprintthis version posted January 4, 2024. ; https://doi.org/10.1101/2024.01.04.572506doi: bioRxiv preprint
With the help of the active learning component highlighting areas of high interest, the experts identified and
added several new behavior types. Figure 4 demonstrates the benefit of using the system to extend the training
dataset.
The novelty behavior detection model was evaluated through a leave-one-out assessment due to the absence of
explicit labels for training and evaluation. This approach involves iteratively treating each known rodent
behavior as novel, allowing us to gauge the proficiency of the method in identifying these established behaviors
as potentially new instances. Further, it gives insights into the ability to expedite the discovery of behaviors by
ranking them higher in the novelty queue. This is crucial for labelers using the labeling assistant algorithm, as it
aids in identifying behaviors promptly rather than randomly later in the process. We observed a significant
improvement in 6 out of 9 behavior types. Some of the behavior types would not have been prioritized by the
novelty ranking, largely due to their intermediate position in the reduced feature space (Figure 3). Thus, we
extended the active learning component to be composed of multiple strategies additionally to novelty ranking.
Classification results
As the amount of behavior types was heavily extended during the project, a direct comparison of model results at
the beginning and the end of the project is not feasible. However, a strong improvement can be seen in behavior
types for which little training data was available at the start of the project due to the rare occurrence of these
behaviors as shown in Table 1. With the data collection using our system and active learning, the amount of
labels for some of these behavior types could be increased substantially. This enabled the training of a model,
which can detect these behaviors more reliably. Examples for this are “Grooming”, where the true positive (TP)
rate improved from 6% to 73%, and “Twitching” with an improvement from 3% to 29%. Moreover, the final
model is able to detect a larger number of distinct behavior types.
Behavior Type
TP Initial Model
TP Final Model
Label Data Increased
Ataxia
0.79
0.83
1.1
Digging
0.02
0.06
4.3
Eating
0.38
0.88
2
Grooming
0.17
0.79
3.88
Sniffing
0.63
0.54
5.64
Startled
0.08
0.09
5.67
Twitching
0.03
0.29
4.8
Unsupported Rearing
0.72
0.72
1.7
Catalepsy
-
0.59
Drinking
-
0.91
Gnawing
-
0.29
Interrupted Sleeping
-
0.45
Jumping
-
0.85
Supported Rearing
-
0.70
Stretched
0.57
Walking
-
0.78
Writhing
-
0.11
Table 1. Comparison of the performance between the initial model and the final model, as well as the difference in data set
size. Each row refers to a class of behavior type that the models aim to detect. TP refers to the true positive rate of the
predictions, evaluated on a per-frame basis. Label Data Increased lists the factor by how much the amount of annotated data
for that class was increased from initial to final training data set. For simplicity, not all classes that the model was trained on
are included.
.CC-BY-NC-ND 4.0 International licensemade available under a
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is
The copyright holder for this preprintthis version posted January 4, 2024. ; https://doi.org/10.1101/2024.01.04.572506doi: bioRxiv preprint
Conclusion
Our study has demonstrated that the integration of a human-in-the-loop approach, combined with advanced AI
technologies, significantly enhances the efficiency and accuracy of rodent behavior analysis in the context of
drug development.
The active learning component has proven instrumental in discovering and annotating rare behavior types, as
evidenced by the substantial increase in annotations and the identification of several new behavior classes. This
enhancement in data richness not only improves the model's accuracy but also broadens the spectrum of behavior
types that can be reliably detected. The increase in the amount of labeled data for rare behaviors has notably
improved the model's performance, as highlighted by the substantial improvements in TP rates for behaviors
such as Grooming and Twitching. The system has been proven to accelerate the discovery process and aid
labelers in prioritizing behaviors for annotation.
DeepRod represents a significant advancement in the field of automated rodent behavior analysis. Its ability to
efficiently process large datasets, coupled with its enhanced detection and classification capabilities, makes it a
powerful tool for drug discovery and development. The collected user feedback confirmed that the speed of
annotating the experiment data and the discovery of such a large amount of new behavior types would not have
been possible without the provided system. As the system continues to evolve, it holds great potential for further
improving the understanding of rodent behavior, contributing to more effective and efficient drug development
processes and safe clinical trials.
Acknowledgements and contributions
Our deepest gratitude goes to Beatrice Kley, Cindy Janine Jung, and Michael Winter (Boehringer Ingelheim
GmbH & Co. KG, Drug Discovery Sciences) for annotation of the data and for their feedback, wishes, ideas and
immediate testing of these. We want to thank Malte Janssen, Moisei Shkil, Ziyad Sheebaelhamd, Lev Udaltsov,
Atacan Korkmaz (Merantix Momentum GmbH) who developed software components used for DeepRod.
DeepRod would not be possible without Florian Montel, Menorca Chaturvedi, and Markus Köster from
opnMe.com [1] (F.M. Boehringer Ingelheim GmbH & Co. KG; M.C., M.K. Boehringer Ingelheim International
GmbH). Fabian Heinemann and Peter Gross contributed the opnMe initiative at various stages and provided
input on the requirements of the MVP. Kathrin Eschmann had a substantial impact on the late stage of the project
and contributed to the manuscript (all Boehringer Ingelheim GmbH & Co. KG, Drug Discovery Sciences).
Adrian Loy, Miha Garafolj, Thomas Wollmann developed the machine learning method. Adrian Loy and Miha
Garafolj did the algorithm implementation and performed the machine learning experiments. Hanna Behnke did
the usability engineering and orchestrated the project. Cyrille Charnier and Philipp Schwarz evaluated the
methodology. Georg Rast, Heike Schauerte contributed the initial concept and provided domain knowledge.
Georg Rast and Thomas Wollmann supervised the project. Thomas Wollmann, Miha Garafolj, Adrian Loy, Heike
Schauerte, Georg Rast, and Hanna Behnke wrote and proofread the manuscript.
Ethical statement
Maintenance and handling of animals are carried out in compliance with (i) the ethical guidelines established by
German National Animal Welfare Laws within the framework of the European Union Directive 2010/63/EU and
(ii) the Guide for the Care and Use of Laboratory Animals produced by the National Research Council and the
Association for Assessment and Accreditation of Laboratory Animal Care International (AAALAC). The study
protocol was approved by the responsible German authority (Regierungspräsidium Tübingen).
.CC-BY-NC-ND 4.0 International licensemade available under a
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is
The copyright holder for this preprintthis version posted January 4, 2024. ; https://doi.org/10.1101/2024.01.04.572506doi: bioRxiv preprint
References:
1. European Parliament and Council (2001). Directive 2001/83/EC of the European Parliament and of the
Council of 6 November 2001 on the Community code relating to medicinal products for human use.
<https://eur-lex.europa.eu/legal-content/DE/TXT/?uri=celex%3A32001L0083>. Accessed 21 December 2013.
2. FDA & EMA, 2001. ICH S7A Safety Pharmacology Studies for Human Pharmaceuticals (2001).
<https://www.fda.gov/regulatory-information/search-fda-guidance-documents/s7a-safety-pharmacology-studies-
human-pharmaceuticals>. Accessed 21 December 2013.
3. Gad SC (2019). Safety Pharmacology in Pharmaceutical Development: Approval and Post Marketing
Surveillance, Second Edition. CRC Press, Chapter 2.1, 19-24 and Chapter 5.1-5.2, 69-77.
4. Hamdam J, Sethu S, Smith T, Alfirevic A, Alhaidari M, Atkinson J, Ayala M, & Goldring C (2013). Safety
pharmacology - current and emerging concepts. Toxicol Appl Pharmacol. 273(2), 229-41.
5. Spink, A.J., Buma, M.O.S., Tegelenbosch, R.A.J. (2000). EthoVision color identification: a new method for
color tracking using both hue and saturation. Proceedings of Measuring Behavior 2000, 295-297.
6. Spink, A.J., Tegelenbosch, R.A.J., Buma, M.O.S., Noldus, L.P.J.J. (2000). The EthoVision video tracking
system: a tool for behavioral phenotyping of transgenic mice. Physiology & Behavior 73, 731-744.
7. Otterbach, J., & Wollmann, T. (2021). Chameleon: A Semi-AutoML framework targeting quick and scalable
development and deployment of production-ready ML systems for SMEs. arXiv preprint arXiv:2105.03669.
8. Mathis, A., Mamidanna, P., Cury, K. M., Abe, T., & Bethge, M. (2018). DeepLabCut: markerless pose
estimation of user-defined body parts with deep learning. Nature Neuroscience, 21, 1281-1289.
9. Cristina Segalin, Jalani Williams, Tomomi Karigo, May Hui, Moriel Zelikowsky, & Ann Kennedy (2021)
The Mouse Action Recognition System (MARS) software pipeline for automated analysis of social behaviors in
mice. eLife 10:e63720.
10. Föll, M. C., Moritz, L., Wollmann, T., Stillger, M. N., Vockert, N., Werner, M., ... & Schilling, O. (2019).
Accessible and reproducible mass spectrometry imaging data analysis in Galaxy. Gigascience, 8(12)
11. Ritter, C., Wollmann, T., Lee, J. Y., Imle, A., Müller, B., ... & Rohr, K. (2021). Data fusion and smoothing for
probabilistic tracking of viral structures in fluorescence microscopy images. Medical Image Analysis, 73.
12. Chen, T., & Guestrin, C. (2016, August). Xgboost: A scalable tree boosting system. Proceedings of ACM
SIGKDD 2016, 785-794.
13. McElfresh, D., Khandagale, S., Valverde, J., Ramakrishnan, G., Goldblum, M., & White, C. (2023). When
Do Neural Nets Outperform Boosted Trees on Tabular Data?. arXiv:2305.02997.
14. Lundberg, S. M., Erion, G., Chen, H., DeGrave, A., Prutkin, J. M., ... & Lee, S. I. (2020). From local
explanations to global understanding with explainable AI for trees. Nature machine intelligence, 2(1), 56-67.
15. Sreenivasaiah, D., Otterbach, J., & Wollmann, T. (2021). Meal: Manifold embedding-based active learning.
Proceedings of IEEE ICCV 2021. 1029-1037.
16. Bitterwolf, J., Müller, M., Hein, M. (2023). In or Out? Fixing ImageNet Out-of-Distribution Detection
Evaluation. Proceedings of ICML 2023.
17. Boehringer Ingelheim (2017). opnMe - Boehringer Ingelheim Open Innovation Portal. <https://opnMe.com>.
Accessed 21 December 2013.
.CC-BY-NC-ND 4.0 International licensemade available under a
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is
The copyright holder for this preprintthis version posted January 4, 2024. ; https://doi.org/10.1101/2024.01.04.572506doi: bioRxiv preprint
ResearchGate has not been able to resolve any citations for this publication.
Article
Full-text available
Tree-based machine learning models such as random forests, decision trees and gradient boosted trees are popular nonlinear predictive models, yet comparatively little attention has been paid to explaining their predictions. Here we improve the interpretability of tree-based models through three main contributions. (1) A polynomial time algorithm to compute optimal explanations based on game theory. (2) A new type of explanation that directly measures local feature interaction effects. (3) A new set of tools for understanding global model structure based on combining many local explanations of each prediction. We apply these tools to three medical machine learning problems and show how combining many high-quality local explanations allows us to represent global structure while retaining local faithfulness to the original model. These tools enable us to (1) identify high-magnitude but low-frequency nonlinear mortality risk factors in the US population, (2) highlight distinct population subgroups with shared risk characteristics, (3) identify nonlinear interaction effects among risk factors for chronic kidney disease and (4) monitor a machine learning model deployed in a hospital by identifying which features are degrading the model’s performance over time. Given the popularity of tree-based machine learning models, these improvements to their interpretability have implications across a broad set of domains. Tree-based machine learning models are widely used in domains such as healthcare, finance and public services. The authors present an explanation method for trees that enables the computation of optimal local explanations for individual predictions, and demonstrate their method on three medical datasets.
Article
Full-text available
Quantifying behavior is crucial for many applications in neuroscience. Videography provides easy methods for the observation and recording of animal behavior in diverse settings, yet extracting particular aspects of a behavior for further analysis can be highly time consuming. In motor control studies, humans or other animals are often marked with reflective markers to assist with computer-based tracking, but markers are intrusive, and the number and location of the markers must be determined a priori. Here we present an efficient method for markerless pose estimation based on transfer learning with deep neural networks that achieves excellent results with minimal training data. We demonstrate the versatility of this framework by tracking various body parts in multiple species across a broad collection of behaviors. Remarkably, even when only a small number of frames are labeled (~200), the algorithm achieves excellent tracking performance on test frames that is comparable to human accuracy.
Article
Full-text available
Video tracking systems enable behavior to be studied in a reliable and consistent way, and over longer time periods than if they are manually recorded. The system takes an analog video signal, digitizes each frame, and analyses the resultant pixels to determine the location of the tracked animals (as well as other data). Calculations are performed on a series of frames to derive a set of quantitative descriptors of the animal's movement. EthoVision (from Noldus Information Technology) is a specific example of such a system, and its functionality that is particularly relevant to transgenic mice studies is described. Key practical aspects of using the EthoVision system are also outlined, including tips about lighting, marking animals, the arena size, and sample rate. Four case studies are presented, illustrating various aspects of the system: (1) The effects of disabling the Munc 18-1 gene were clearly shown using the straightforward measure of how long the mice took to enter a zone in an open field. (2) Differences in exploratory behavior between short and long attack latency mice strains were quantified by measuring the time spent in inner and outer zones of an open field. (3) Mice with hypomorphic CREB alleles were shown to perform less well in a water maze, but this was only clear when a range of different variables were calculated from their tracks. (4) Mice with the trkB receptor knocked out in the forebrain also performed poorly in a water maze, and it was immediately apparent from examining plots of the tracks that this was due to thigmotaxis. Some of the latest technological developments and possible future directions for video tracking systems are briefly discussed.
Article
Automatic tracking of viral structures displayed as small spots in fluorescence microscopy images is an important task to determine quantitative information about cellular processes. We introduce a novel probabilistic approach for tracking multiple particles based on multi-sensor data fusion and Bayesian smoothing methods. The approach exploits multiple measurements as in a particle filter, both detection-based measurements and prediction-based measurements from a Kalman filter using probabilistic data association with elliptical sampling. Compared to previous probabilistic tracking methods, our approach exploits separate uncertainties for the detection-based and prediction-based measurements, and integrates them by a sequential multi-sensor data fusion method. In addition, information from both past and future time points is taken into account by a Bayesian smoothing method in conjunction with the covariance intersection algorithm for data fusion. Also, motion information based on displacements is used to improve correspondence finding. Our approach has been evaluated on data of the Particle Tracking Challenge and yielded state-of-the-art results or outperformed previous approaches. We also applied our approach to challenging time-lapse fluorescence microscopy data of human immunodeficiency virus type 1 and hepatitis C virus proteins acquired with different types of microscopes and spatial-temporal resolutions. It turned out, that our approach outperforms existing methods.
Book
Safety pharmacology is the evaluation and study of the pharmacological effects of a potential drug that are unrelated to the desired therapeutic effect. These effects often present a hazard-particularly in individuals with compromised or limited organ system functions. Safety Pharmacology in Pharmaceutical Development: Approval and Post Marketing Surveillance, Second Edition covers safety pharmacology from the regulatory requirements down to the studies that must be done to justify them. Using the author's more than 30 years of direct experience, the book incorporates tricks and practical insights for making studies work and understanding why they fail. The second edition includes current regulations, including USFDA and those from Europe and Japan. Presenting a clear description of what is needed and why for supporting drug development, the book focuses on updated test methods, interpretation, and science. It covers the core and supplemental batteries of test procedures and how to do them and provides an overview of available facilities and contract organizations for performing studies.
Conference Paper
Tree boosting is a highly effective and widely used machine learning method. In this paper, we describe a scalable end-to-end tree boosting system called XGBoost, which is used widely by data scientists to achieve state-of-the-art results on many machine learning challenges. We propose a novel sparsity-aware algorithm for sparse data and weighted quantile sketch for approximate tree learning. More importantly, we provide insights on cache access patterns, data compression and sharding to build a scalable tree boosting system. By combining these insights, XGBoost scales beyond billions of examples using far fewer resources than existing systems.
Directive 2001/83/EC of the European Parliament and of the Council of 6
  • European Parliament
  • Council
European Parliament and Council (2001). Directive 2001/83/EC of the European Parliament and of the Council of 6 November 2001 on the Community code relating to medicinal products for human use. <https://eur-lex.europa.eu/legal-content/DE/TXT/?uri=celex%3A32001L0083>. Accessed 21 December 2013.
EthoVision color identification: a new method for color tracking using both hue and saturation
  • A J Spink
  • M O S Buma
  • R A J Tegelenbosch
Spink, A.J., Buma, M.O.S., Tegelenbosch, R.A.J. (2000). EthoVision color identification: a new method for color tracking using both hue and saturation. Proceedings of Measuring Behavior 2000, 295-297.
Chameleon: A Semi-AutoML framework targeting quick and scalable development and deployment of production-ready ML systems for SMEs
  • J Otterbach
  • T Wollmann
Otterbach, J., & Wollmann, T. (2021). Chameleon: A Semi-AutoML framework targeting quick and scalable development and deployment of production-ready ML systems for SMEs. arXiv preprint arXiv:2105.03669.