ChapterPDF Available

Comparative Study of Machine Learning Methods for In-Vehicle Intrusion Detection

Authors:
  • Independent Researcher
  • St. Petersburg Federal Research Center of the Russian Academy of Sciences (SPC RAS)

Abstract and Figures

An increasing amount of cyber-physical systems within modern cars, such as sensors, actuators, and their electronic control units are connected by in-vehicle networks and these in turn are connected to the evolving Internet of vehicles in order to provide “smart” features such as automatic driving assistance. The controller area network bus is commonly used to exchange data between different components of the vehicle, including safety critical systems as well as infotainment. As every connected controller broadcasts its data on this bus it is very susceptible to intrusion attacks which are enabled by the high interconnectivity and can be executed remotely using the Internet connection. This paper aims to evaluate relatively simple machine learning methods as well as deep learning methods and develop adaptations to the automotive domain in order to determine the validity of the observed data stream and identify potential security threats.
Content may be subject to copyright.
Sokratis K. Katsikas · Frédéric Cuppens
Nora Cuppens · Costas Lambrinoudakis
Annie Antón · Stefanos Gritzalis
John Mylopoulos · Christos Kalloniatis (Eds.)
123
LNCS 11387
ESORICS 2018 International Workshops,
CyberICPS 2018 and SECPRE 2018
Barcelona, Spain, September 6–7, 2018
Revised Selected Papers
Computer Security
Comparative Study of Machine Learning
Methods for In-Vehicle Intrusion
Detection
Ivo Berger1,RolandRieke
2,3(B
),MaximKolomeets
3,4 ,
Andrey Chechulin3,4 ,andIgorKotenko
3
1TU Darmstadt, Darmstadt, Germany
2Fraunhofer Institute SIT, Darmstadt, Germany
roland.rieke@sit.fraunhofer.de
3ITMO University, St. Petersburg, Russia
4SPIIRAS, St. Petersburg, Russia
Abstract. An increasing amount of cyber-physical systems within mod-
ern cars, such as sensors, actuators, and their electronic control units are
connected by in-vehicle networks and these in turn are connected to
the evolving Internet of vehicles in order to provide “smart” features
such as automatic driving assistance. The controller area network bus is
commonly used to exchange data between dierent components of the
vehicle, including safety critical systems as well as infotainment. As every
connected controller broadcasts its data on this bus it is very susceptible
to intrusion attacks which are enabled by the high interconnectivity and
can be executed remotely using the Internet connection. This paper aims
to evaluate relatively simple machine learning methods as well as deep
learning methods and develop adaptations to the automotive domain in
order to determine the validity of the observed data stream and identify
potential security threats.
Keywords: Machine learning ·Automotive security ·
Internet of vehicles ·Predictive security analysis ·
System behavior analysis ·Security monitoring ·Intrusion detection ·
Controller area network security
1Introduction
Each modern vehicle can be regarded as a system of interconnected cyber-
physical systems. When vehicles are to take over tasks which are up to now the
responsibility of the driver, an increasing automation and networking of these
vehicles with each other and with the infrastructure is necessary. In particular,
autonomous driving requires both a strong interconnectedness of vehicles and an
opening to external information sources and services, which increases the poten-
tial attack surface. An indispensable assumption, however, is that the vehicle can
not be controlled unauthorized externally. Thus, IT security and data protection
c
!Springer Nature Switzerland AG 2019
S. K. Katsikas et al. (Eds.): CyberICPS 2018/SECPRE 2018, LNCS 11387, pp. 85–101, 2019.
https://doi.org/10.1007/978-3-030-12786-2_6
86 I. Berger et al.
are enabling factors for the newly emerging Internet of Vehicles (IoV). Practical
experiments [17] have already demonstrated that an attacker can gain remote
access to an in-vehicle Electronic Control Unit (ECU) and recent advisories such
as [10]alsomentionthatpublicexploitsareavailable.
This work now examines various methods by which activities of an attacker
already inside a vehicle could be detected. The Controller Area Network (CAN)
bus is the standard solution for in-vehicle communication between the ECUs of
the in-vehicle cyber-physical systems. Whilst oering high reliability the CAN
bus lacks any kind of built-in security measures to prevent malicious interference
by an outside party. This makes it easy for an attacker with access to one ECU
to take over other critical cyber-physical systems within a vehicle. This could
be done by broadcasting forged commands on the network, to gain the required
knowledge, e.g. by running a fuzzing attack or to impair bus performance by
performing a simple Denial of Service (DoS) attack.
In this work, we implemented and analyzed anomaly detection methods which
can be applied to existing vehicle architectures as well as new designs by a soft-
ware update or plug-in module as proposed in [19]withouttheadoptionofnew
communication standards. The main contribution of this work is a comparative
assessment of dierent Machine Learning (ML) methods with respect to usability
for intrusion and anomaly detection within automotive CAN networks.
Section 2introduces data sets from two dierent cars with and without
attacks that have been used to evaluate the compared methods. Section 3pro-
vides the background on four dierent ML technologies used in this work includ-
ing training, validation and performance assessments. Section4presents the
results of various experimental setups, while Sect.5discusses these outcomes
and gives recommendations on feasible approaches for the domain of in-vehicle
networks. Section 6describes related work, and Sect. 7concludes this paper.
2 Data Sets
On the CAN bus every attached device broadcasts messages. At the same time,
devices listen for relevant information. We mapped the relevant data of CAN
messages to the following tuple structure: (time, ID, dlc, p1,...,p
8,type), where
time is the time when the message was received, ID comprises information
about the type and the priority of the message, dlc (data length code) provides
the number of bytes of valid payload data, p1,...,p8isthepayloadof08
bytes, and type marks the message (attack versus no attack). In cases where
dlc < 8 we inserted dummy content in order to have a fixed tuple structure.
For the experiments in this work we used five dierent data sets, namely, ZOE,
HCLRDoS,HCLRFuzzy,HCLRRP M , and HCLRGear . The ZOE data set has
been collected from a 10 min drive with a Renault Zoe electric car in an urban
environment. It contains about 1 million messages and has been used before
in [22] to perform behavior analysis by process mining. This data set contains
no attack data. The other four data sets that we used have been published by
the “Hacking and Countermeasures Research Labs” (HCRL) [7]. These data
Comparative Study of Machine Learning Methods 87
sets are fully labeled and demonstrate dierent attack types. The HCLRDoS
data set contains DoS attacks. For this attack, every 0.3 ms a message with the
ID “0000” is injected. Conversely, in the HCLRFuzzy data set every 0.5 ms a
completely random message is injected, whereas HCLRRPM and HCLRGear
contain spoofing attacks. In these data sets every millisecond a message with an
ID related to gear respectively engine revolutions per minute is injected. The ID
and message does not change. The linear charts of the ZOE,HCLRDoS, and
HCLRGear data sets depicted in Fig.1show the composition of legal data and
attacks. Figure 1ashowsthatintheHCLRDoS data set DoS messages decrease
the number of legal packets, whereas Fig. 1b unveils a big gap in the trac time-
line of the HCLRGear data set which could probably be a consequence of the
spoofing attack. The linear charts of HCLRFuzzy and HCLRRP M not shown
here are similar to HCLRGear.
(a) HC LRDoS : DoS attacks (b) H C LRGear : spoofing
Fig. 1. Attack influence analysis by linear graphs
In order to get some more insight into the contents of the CAN data sets, we
have visualized them by radial time-intervals using the method described in [12].
Figure 2shows dierences between ZOE and HCLRDoS trac. The significantly
higher number of bars in Fig. 2a is due to a higher number of dierent IDs in
the ZOE trac. In Fig. 2a the trac without attacks has no outstanding bars
whereas in Fig. 2bsolidorangebarsareoutstandingincomparisontoother
bars. These bars represent DoS attacks which are decreasing the number of legal
messages in the first three intervals during the attack. The radial visualizations
of the other HCRL data sets with fuzzing and spoofing injections not shown here
did not provide more insights.
3MachineLearningMethods
This section introduces the compared algorithms including training, validation
and performance assessments. The problem of detecting anomalies and attacks
in CAN data diers from most ML applications as it does not present a clear
88 I. Berger et al.
Fig. 2. Attack influence visualization by radial time intervals, each summarizing one
quarter of the period represented by the data set; trac is separated in four radial time
intervals that consist of bars; each ID is represented as a bar whose height equals the
number of messages; bars consists of arcs which represent payload the more messages
with same payload the higher is the arc; so solid (or almost solid) arcs depict messages
with few dierent payloads; transparent bars depict messages with big payload variety.
(Color figure online)
classification problem. Anomalies and attacks are by nature unpredictable and
thus it is not possible to obtain representative data to train a classifier. Thus,
the approach taken in this work is to fit a model for the regular system behavior
which can detect deviations from the norm. One aspect of system behavior is
the range of values for the IDs and the payload of CAN messages, another
aspect is the temporal behavior, resulting in a novelty detection problem and a
time series analysis task. Some ECUs send messages periodically and thus are
relatively easy to validate. Even though the collection of a representative set of
anomalies and attacks is not possible it is beneficial to be able to detect and
prevent known types of attacks such as DoS or fuzzing attacks. This presents
a standard classification task where a model is trained to dierentiate between
regular and anomalous CAN communications. Another aspect is the practicality
of possible solutions in real-world scenarios. For the deployment in vehicles the
trained models need to be able to validate the incoming data steam in real-time,
requiring ecient models and thus restricting their complexity.
The training of all models was done using the data sets introduced in Sect. 2
or subsets thereof in order to achieve reasonable training times. The data was
split into training and validation sets to get a realistic performance estimate for
each model. The validation of the Support Vector Machine (SVM) and standard
neural network models use accuracy and confusion matrices. The Long Short
Term Memory (LSTM) network was validated by predicting the next message ID
Comparative Study of Machine Learning Methods 89
based on a window of preceding messages and comparing it to the actual message.
All experiments in this paper are written in Python 3.6.Theyutilizethepandas
[16] library to read and transform the data and scikit-learn [21] and keras [6]
with the tensorflow [1] back-end for the ML itself. For visualization matplotlib
[9] and seaborn [28] were used. We now give a short introduction on how each
of the methods operate.
3.1 One-Class Support Vector Machines
For anomaly detection One-Class Support Vector Machines (OCSVM) were used,
which are an adaption of classic SVMs to be trained with one class with the goal
of learning a border which encompasses the training data. OCVSMs are linear
classifiers but can make use of nonlinear kernels to represent more complex data
structures. They were used with success in [4,27] and this work used the hyper-
parameters suggested in [4]. For OCSVMs the sklearn.svm.OneClassSVM and
numpy [20] packages were used to filter out anomalous data from the training set.
The scikit-learn [21]metricsaccuracy score and confusion matrix were
used to calculate scores from the predictions on the test set. To visualize the
results the metrics for all data sets were saved and displayed using a seaborn
[28]heat-mapfortheconfusionmatrices(Fig.4)andasimplelinegraphforthe
accuracy per subset size (Fig. 3).
3.2 Support Vector Machines
SVMs are linear “max-margin” classifiers as they try to find a hyper-plane sep-
arating the data with the greatest possible margin to the closest entry of each
class. They are linear but can use kernels to model nonlinear data structures
whilst maintaining low hardware requirements when classifying. As they are very
similar to OCSVMs the hyper-parameters from [4] were used here as well. Our
implementation of the regular SVMs is almost identical to the OCSVM, except
that sklearn.svm.NuSVC was used instead of sklearn.svm.OneClassSVM.
3.3 Neural Networks
Neural networks are the standard for deep-learning and can model very complex
nonlinear relationships. The most basic version is the fully connected neural
network. It utilizes an arbitrary number of layers with each layer supporting an
arbitrary number of neurons. Data is propagated from the input to the output
layer using weighted connections between the neurons of these layers, resulting
in very complex structures and thus a large amount of trainable parameters and
thus flexibility even for relatively small networks. They are usually trained using
some form of gradient descent and are prone to overfitting due to their great
flexibility. In consequence, the goal was to find the smallest possible network to
achieve a good accuracy. Therefore, the anomalous class was set to 0 to work
properly for binary classification and all features of the complete set were scaled
using the MinMaxScaler from scikit-learn [21]beforetraining.
90 I. Berger et al.
The neural networks were implemented using keras [6]Sequential model
from the keras.models package and keras.layers.Dense as its fully connected
layers. From initial test it was found that one hidden layer and one epoch is
sucient for these data sets. For easier testing both layers used the same number
of neurons. Binary Crossentropy, Adam and Accuracy was used as loss, optimizer
and performance metric (see Listing 1.1).
Listing 1.1. Neural Network: Model and Training
def train (x: np .ndarray , y: np.ndarray , split , batch size ,
neurons ) :
#define and compile the model
model = Sequen tia l ()
model . add (Dens e( n eurons , a c ti v at io n= r el u , in put shape=(x
.shape[1] ,)))
model . add( Dense ( neu ron s ) )
model . add ( Dense (1 , ac t iv a ti o n=’ s igm oid ) )
model . c omp ile ( l o ss= bi na ry crossentropy , optimizer=’adam’
,metrics=[accuracy])
#train model
model . f i t ( x , y , ep ochs =1, b at ch size=batch size , shuffle=
True , ve rbose =0)
return model
Due to the good optimization of the tensorflow [1] back-end the model
wasn’t tested with dierent subset sizes but dierent neuron counts.
3.4 Long Short Term Memory Neural Networks
LSTMs are a derivation of recurrent neural networks for time sequence classi-
fication and prediction. They dier from standard neural networks by keeping
previous states and thus are able to capture temporal relationships. LSTMs in
particular keep a very recent as well as a long-standing state and are able to
detect relationships between relatively distant events as well as directly consec-
utive ones opposed to simpler recurrent neural networks which only remember
recent states. LSTMs are trained with time sequences, requiring to pre-process
the data sets into message sequences with the window size as a configurable
parameter. Furthermore, they are not trained to classify a message as anoma-
lous or non-anomalous but to predict future messages or validate if new messages
concur with the learned behavior.
Training a L S T M t o p re d i ct o r validat e n e w m es s a ge s r e q ui r e s th e t i me s e r ie s
that is the training data to be transformed into a supervised learning problem.
This is achieved by using message sequences of a certain window size with the
message ID that followed it instead of single messages with a binary label. This
enables the LSTM to learn the behavior and temporal relationships between data
points. The original version of the code used was taken from [2]andhasbeen
adapted and simplified for this work. The next pre-processing step is the trans-
formation of the time stamps to time deltas per ID, i.e. that the time column
Comparative Study of Machine Learning Methods 91
gives the seconds since the last occurrence of that ID instead of a compara-
tively arbitrary time stamp. This is done using pandas [16] split-apply-combine
pandas.DataFrame.grouby and apply functions. To make the problem easier to
solve and thus the training times shorter the predictions were limited to the mes-
sage IDs instead of predicting/validating whole messages. To achieve good results
the IDs had to be transformed from simple numbers to categories keras [6] and
tensorflow [1] can properly handle. This process utilizes the scikit-learn [21]
LabelEncoder and keras [6]to categorical functions which first encode the
IDs as labels and then transform them into a one-hot encoded numpy [20]array.
The last step before applying the time series transformation is a MinMaxScaler.
For usage with keras [6]theresultofthetimeseriestransformationhastobe
reshaped into a three-dimensional array containing the original data point, the
following ten steps and the corresponding label.
Listing 1.2. LSTM: Training
def train(x, y , batch size , neurons=10):
#define and compile the model
model = Sequen tia l ()
model . add (LSTM( ne urons , in pu t shape=(x . shape [1 ] , x. shape
[2])))
model . add (Dense ( y . s hape [ 1 ] , a ct i va t io n=’ so ftma x ) )
model . c omp ile ( l o ss= c at e go ri c al crossentropy ,
optimizer=’adam , metrics=[ ’accuracy ])
#train the model
model . f i t ( x , y , ep ochs =5, b at ch size=batch size , shuffle=
False , verbose=0)
return model
Listing 1.2 shows the actual training process which is quite similar to that of
aregularneuralnetwork.Dierencesareintheusedlossfunction(categoricalvs.
binary crossentropy) and that the data isn’t shued for LSTMs as that would
destroy any temporal relationships in the data.
The scoring is essentially the same as for the neural networks with the excep-
tion that the predictions are given as probabilities per category which have to
be transformed to the one-hot encoding in the test set by setting the category
with the highest probability to one and all others to zero.
4Results
This section presents the results of all methods mentioned in Sect. 3.Itwill
introduce the metrics used and discuss the performance of each method with
regard to the nature of the data sets and validation methods.
4.1 One-Class Support Vector Machines
The OCSVMs were validated with subsets of the data sets described in Sect. 2
of dierent sizes between 5,000 and 300,000 messages using two dierent
92 I. Berger et al.
approaches: The ZOE data set consists of non-anomalous data only, result-
ing in a validation error that is equivalent to the false negative rate, i.e. it was
tested which percentage of the validation data was misclassified as anomalous.
The training portion of the HCRL data set was cleaned of anomalous data and
the trained model tested with both classes using the accuracy for performance
assessment as well as confusion matrices where appropriate with the true label
on the y-axis and the predicted label on the x-axis.
(a) OCSVM with time-stamps (b) OCSVM without time-stamps
Fig. 3. OCSVM results
The high accuracy on the HCLRDoS data set is expected as the anomalous
entries are easily detectable for any subset of the data. Figure3a, however, shows
aslightworseningwithincreasingsubsetsizes.Asthemodelwastrainedwithall
fields, including the timestamps, the result suggest that the OCSVM is detect-
ing messages with a timestamp outside of the learned boundaries as anomalous.
Excluding the timestamps from training and testing confirms this as it results
in almost perfect accuracy for all subset sizes (see Fig. 3b). As seen in Fig. 3, the
OCSVMs accuracy on the ZOE data is almost perfect. Considering that this
data set only consists of regular data the good result comes from a too great
similarity of the training and test data sets. The result thus lacks informative
value about the eectiveness of OCSVMs in anomaly detection. The performance
on the HCLRFuzzy data set is pretty high on a subset of 50,000 messages and
declines with increasing message count. This can be explained with the random-
ized generation of the anomalous data in this set. With increasing subsets the
amount of completely random data increases as well, which in turn increases the
amount of anomalous data that looks like regular data by chance, resulting in
deterioration of the results. This is confirmed by the confusion matrix in Fig. 4.
The model predicts the regular class almost exclusively and the performance
changes are a result of changes in the test set rather than changes in the model.
Comparative Study of Machine Learning Methods 93
The accuracy for the spoofing data sets first shows a slight dip for 50,000 sam-
ples and then recovers with larger subsets. The confusion matrices in Fig.4show
that this is purely due to changes in the test data set as the model only predicts
the regular class. The anomalous data in these sets only dier in the timestamp.
There is nothing in the ID or payload of these entries that makes them distinct
from other non-anomalous messages, making it impossible for a SVM to sepa-
rate between classes. The confusion matrices (Fig. 4)showheavybiastowards
the regular class for all data sets except HCLRDoS and thus that all changes
in the performance of the models are due to changes in the composition of the
test data and not an improvements of the models itself. We also observed that
the removal of the timestamps only has a noticeable eect on the HCLRDoS
results. This can be explained with the mentioned bias as well as any potential
changes are shadowed by the almost exclusive prediction of the regular class for
all other data sets.
Fig. 4. OCSVM confusion matrices
4.2 Support Vector Machines
SVMs are very similar to OCSVMs as described in Sect. 3.2,hencethemethod
of validation is as described for OCSVMs in Sect. 4.1 with the important dier-
ence that SVMs are classifiers and thus were trained with regular and anomalous
entries. As SVMs don’t support training on data sets with only one class the
ZOE data set is excluded from these results. Furthermore, the SVM implemen-
tation in scikit-learn [21] is not multithreaded and had very long training
times when training with timestamps. For these reasons only results without
timestamps are presented. The results in Fig. 5ashowthatknowledgeaboutthe
anomalous entries significantly improves accuracy on the impersonation data
sets, achieving perfect classification on all but the HCLRFuzzy data set even
with the smallest subset. Looking at the very obvious distinction between regu-
lar and anomalous data points in these sets (see Sect. 2) the good performance
is as expected as the continuously worsening performance on the HCLRFuzzy
data set.
94 I. Berger et al.
(a) SVM without timestamps (b) SVM confusion matrix
Fig. 5. Results from support vector machines
Considering that all anomalous entries for this set are random and the result-
ing possibility of entries falling within the value range of regular entries there are
no support vectors that can describe the dierence comprehensively. Thus, the
accuracy declines with increasing subset size as more and more false negatives
are introduced which is shown in the corresponding confusion matrix in Fig.5b.
Despite the clear decline the classification is still surprisingly good considering
the simple linear kernel and the random nature of the anomalous entries.
4.3 Neural Networks
The validation of the neural network was done with a standard train/test split
of the original data and the performance of networks with dierent amounts of
neurons compared in order to find the simplest possible network to solve the
problem. As neural networks are very flexible and even small ones already have
a good amount of variables this paper examines a network with only one hidden
layer for neurons counts of 2, 5, 10 and 20, going from extremely simple to fairly
complex models. The deep learning results are presented as confidence intervals
which are obtained using the bootstrap method with 50 iterations and a sample
size of 800.000 per iteration.
The very good to perfect results in Fig. 6for all used data sets show the
great flexibility of neural networks. For all data sets the intervals reach 99%
even for the simplest network. The explanation for the good performance can be
found in the very simple structure of the anomalous entries for the HCLRDoS,
HCLRGear and HCLRRP M data sets: in each case there is one exact value
combination that has to be detected. Whilst the OCSVMs had problems (cf.
Sect. 4.1)withtheHCLRGear and HCLRRP M data sets as their anomalous
entries values are within range of regular ones a neural network can learn to single
out this exact combination as being anomalous and thus achieve the seen results.
The intervals and the outliers in particular show that the networks performance
depends greatly on the samples used.
Comparative Study of Machine Learning Methods 95
Fig. 6. Neural network confidence intervals
The case where the very good results are not as obvious is the HCLRFuzzy
data set, as it has randomly generated anomalous entries which can not be
as easily dierentiated from the regular ones, which is supported by the need
of at least 5 neurons to surpass the 99% accuracy threshold. In this case the
great flexibility of the network enables it to learn which value combinations, for
example in which range an ECU’s payloads are, are valid and thus to distinguish
them from the random entries very well. Another observation is that the intervals
are generally larger then for the other data sets. As we didn’t use stratified splits
this suggests that a certain minimum of both regular and random data is required
for the network to learn a good model.
The outliers seen in most results illustrate the general importance of having
the right data. Out results show that with a proper amount of data to train the
model neural networks are capable of detecting complex anomalies reliably.
96 I. Berger et al.
4.4 Long Short Term Memory Neural Networks
To vali d a t e t he L S T Ms t h e w ho l e d a ta s e t w a s t r a ns f o rm e d t o me s s ag e s e q ue n c es
as explained in Sect. 3.4 from which the first 80% were used to train bootstrap
networks while the remaining 20% validated the message ID predictions the
LSTM made. The results graphs (Fig. 7)showthenetworksperformancefor
awindowsizeof10andeachindividualgraphplotscondenceintervalson
accuracy against neuron count of the LSTM layer. The LSTMs were trained and
tested without anomalous entries in order to measure their capability to model
the regular data stream.
Fig. 7. LSTM confidence intervals
For the HCLRDoS and the spoofing data sets the accuracy is quite good con-
sidering the relatively small samples and networks used as well as the complexity
of the problem. On all of these data sets the network achieves an accuracy up to
50% with only 5 neurons and up to 60% on the DoS data set with 20 neurons.
The very similar and good performance is due to equal and relatively low number
of regular message IDs in the sets at only 26. Performance on the HCLRFuzzy
Comparative Study of Machine Learning Methods 97
data set is noticeably but not significantly worse for the simplest LSTM and very
similar to the other HCRL sets for 10 or more neurons. The lower accuracy on
the LSTM with 5 neurons is due to the higher number of IDs at 38. The ZOE
data set with its considerably higher number of message IDs at 110 performs by
far the worst at a maximum accuracy just above 40%. Considering that random
performance for 110 IDs is at only 0.9% an accuracy over 40% is more than 42
times better than random and thus still quite good. As the data set is from a
dierent source than the others (see Sect. 2)itsuggestthattheRenaultZoehas
more complex internals than the vehicles used to acquire the HCRL data sets.
For all tested data sets there is a clear correlation between message ID fre-
quency and the LSTMs categorical accuracy on that ID. The periodic occurrence
of these highly frequent IDs and the possible triggering of reactions to certain
periodic IDs explains the good predictions. Consequently, the worst performance
can be observed on infrequently occurring IDs, especially as they contain IDs
triggered by outside events and thus are simply unpredictable.
5Discussion
The OCSVM results show very clearly that this comparatively simple method
of novelty detection only works for very basic anomaly detection. As OCSVMs
try to find a boundary which contains all or most of the seen data it can only
detect anomalies which dier significantly from the normal data in terms of raw
field values. Considering the observed heavy bias towards the regular class it can
still be useful: if it does classify a message as being anomalous there is a high
chance that it’s correct. Theissler [27] has also conducted a more sophisticated
approach. He used Support Vector Data Descriptors, a derivation of OCSVMs,
trained with message sequences instead of individual messages with better results
and very low to no false anomalies. The low false negative rate and the ability to
train them without anomalous data is a quite important aspect. Combined with
their relatively simple and thus fast classification makes them a good practical
choice for real-time classification in a vehicle.
Regular SVMs share the good results and speed of OCSVMs but require
anomalous data for training. Practical use would only come from the classifica-
tion of attacks which cannot be easily specified such as the fuzzing attack. For
any simple specification violating these specifications could be used directly to
verify the data stream without the need to train a model. Because of that and
the significant decline on performance of SVMs on the HCLRFuzzy data set the
real-world applicability of SVMs for the here evaluated use case is very limited.
Neural networks share the major drawback of needing anomalous training
data to be of any use but show impressive performances on all tested data sets.
The results suggest that they are able to learn the ECU behavior very well
and thus detect diverging data points. Even on the randomly generated fuzzing
attack their accuracy was close to perfect. Considering that only a very simple
network with one hidden layer and two neurons per layer is needed to achieve this
performance and thus classifying very fast, even without a graphics processing
98 I. Berger et al.
unit, they could be a powerful tool to simplify automated specification checking.
In practice, they can be trained with regular and randomly generated data and
automatically derive specifications for non-anomalous data. This would require
additional testing on more diverse data sets in order to generalize this approach.
LSTMs are by far the most complex and thus computationally intensive of
the here presented learning algorithms. Our results show their ability to learn
the behavior at least partially and they have been applied to the more dicult
problem of prediction complete messages with success in [4]. The performance
shows diminishing returns when using more than 20 neurons and further simpli-
fication might be possible by excluding messages triggered by external events.
This opens the possibility of improving the network’s performance while reduc-
ing its complexity as in the present experiments the accuracy is clearly linked to
the number of message IDs. Considering all of the above points LSTMs present
a practical and potentially the most powerful approach of anomaly detection out
of the methods analyzed in this work.
It is interesting to note that none of the ML methods indicated the big gaps
in some of the data sets found by the visualization technique (cf. Fig.1b).
6 Related Work
A collection of possible intrusion points together with proposals for countermea-
sures such as cryptography, anomaly detection and ensuring software integrity
by separating critical systems are presented in [25,30]. Whilst the proposed mea-
sures should prove eective most of them require hardware changes, conflicting
with backwards-compatibility. CAN intrusion detection methods can be grouped
into four categories: (1) Work on detection of ECU impersonating attacks such
as [3,5] in most cases uses some kind of physical fingerprinting by voltage or
timing analysis with specific hardware. This work seeks to mitigate the gen-
eral problems of missing authenticity measures in CAN bus design and thus
is complementary to the work presented in this paper. (2) Work on detection
of specification violations assumes that the specification of normal behavior is
available and thus there is an advantage that no alerts based on false positives
will be generated. The specification based intrusion detection can use specific
checks, e.g. for formality, protocol and data range [13], a specific frequency sen-
sor [8], a set of network based detection sensors [18], or specifications of the state
machines [24]. (3) Work on detection of message insertions can be based on var-
ious technologies like analysis of time intervals of messages [23] or LSTM [29].
(4) Work on detection of sequence context anomalies comprises process min-
ing [22], hidden Markov models [14,19], OCSVM [27], neural networks [11], and
detection of anomalous patterns in a transition matrix [15]. In most cases the
authors of the above mentioned papers described experiments with one specific
method. However, because the authors use dierent data sets for their experi-
ments the results of their work are not directly comparable.
Therefore, we compared dierent ML algorithms with the same data sets. The
closest work to our paper is [4] and [26] which also provides results on method
Comparative Study of Machine Learning Methods 99
comparison. OCSVM, Self Organizing Maps, and LSTM are used in [4] and
LSTM, Gated Recurrent Units (GRU), as well as Markov models are used in [26].
However, in [4] only one small training set of 150,000 packets from the Michigan
Solar Car Team was used, and [26]ismorefocusedonattackgeneration.
7Conclusion
In conclusion this study has shown the potential of ML for anomaly detection
in CAN bus data. Even simple linear methods like OCSVMs can yield good
results when detecting simple attacks while more complex neural networks are
capable to learn “normal” message content from CAN data. The most sophisti-
cated models, namely LSTMs, are able to learn ECU behavior adequately. Even
the deep learning approaches can be kept relatively simple meaning all analyzed
methods should be able to detect anomalies in real-time even on low-end hard-
ware. Combined with the existing research ML promises to be an eective way
to increase vehicle security. The injected attacks are relatively trivial in nature
requiring additional research with more diverse and complex intrusions as well
as the comparison of methods used in other research to the here present ones.
Focused tests, potentially in cooperation with vehicle manufacturers, have to
provide further insights in the prediction capabilities of LSTMs. Furthermore,
real-world tests on practical hardware are needed to confirm that real-time detec-
tion is indeed possible. Based on reliable anomaly detection, appropriate reac-
tions such as simple notifications or automated prevention measures need to be
investigated.
Acknowledgements. This research is partially supported by the German Federal
Ministry of Education and Research in the context of the project secUnity (ID
16KIS0398) and by the Government of Russian Federation (Grant 08-08), by the bud-
get (project No. AAAA-A16-116033110102-5).
References
1. Abadi, M., Agarwal, A., Barham, P., et al.: TensorFlow: large-scale machine learn-
ing on heterogeneous distributed systems. In: 12th USENIX Symposium on Oper-
ating Systems Design and Implementation, OSDI 2016, pp. 265–284 (2016)
2. Brownlee, J.: How to convert a time series to a supervised learning problem
in Python. https://machinelearningmastery.com/convert- time- series- supervised-
learning-problem-python/ (2018). Accessed 28 June 2018
3. Cho, K., Shin, K.G.: Fingerprinting electronic control units for vehicle intru-
sion detection. In: Holz, T., Savage, S. (eds.) 25th USENIX Security Symposium,
USENIX Security 16, 10–12 August 2016, Austin, TX, USA, pp. 911–927. USENIX
Association (2016)
4. Chockalingam, V., Larson, I., Lin, D., Nofzinger, S.: Detecting attacks on the CAN
protocol with machine learning (2016)
5. Choi, W., Joo, K., Jo, H.J., Park, M.C., Lee, D.H.: Voltageids: low-level communi-
cation characteristics for automotive intrusion detection system. IEEE Trans. Inf.
Forensics Secur. 13(8), 2114–2129 (2018)
100 I. Berger et al.
6. Chollet, F., et al.: Keras. https://github.com/keras-team/keras (2015)
7. Hacking and Countermeasure Research Lab (HCRL): Car-hacking dataset for the
intrusion detection. http://ocslab.hksecurity.net/Datasets/CAN- intrusion-dataset
(2018). Accessed 28 June 2018
8. Hoppe, T., Kiltz, S., Dittmann, J.: Security threats to automotive CAN networks
- practical examples and selected short-term countermeasures. Reliab. Eng. Syst.
Saf. 96, 11–25 (2011)
9. Hunter, J.D.: Matplotlib: a 2D graphics environment. Comput. Sci. Eng. 9(3),
99–104 (2007)
10. ICS-CERT: Advisory (ICSA-17-208-01). https://ics-cert.us- cert.gov/advisories/
ICSA-17-208-01, July 2017. Accessed 17 Sept 2018
11. Kang, M.J., Kang, J.W.: A novel intrusion detection method using deep neural
network for in-vehicle network security. In: 2016 IEEE 83rd Vehicular Technology
Conference (VTC Spring) (2016)
12. Kolomeets, M., Chechulin, A., Kotenko, I.: Visual analysis of CAN bus trac
injection using radial bar charts. In: Proceedings of the 1st IEEE International
Conference on Industrial Cyber-Physical Systems, ICPS-2018, Saint-Petersburg,
Russia, pp. 841–846. IEEE (2018)
13. Larson, U.E., Nilsson, D.K., Jonsson, E.: An approach to specification-based attack
detection for in-vehicle networks. In: 2008 IEEE on Intelligent Vehicles Symposium,
pp. 220–225, June 2008
14. Levi, M., Allouche, Y., Kontorovich, A.: Advanced analytics for connected cars
cyber security. CoRR abs/1711.01939 (2017)
15. Marchetti, M., Stabili, D.: Anomaly detection of CAN bus messages through anal-
ysis of ID sequences. In: 2017 IEEE Intelligent Vehicles Symposium (IV), pp. 1577–
1583, June 2017
16. McKinney, W.: Data structures for statistical computing in Python. In: Proceed-
ings of the 9th Python in Science Conference 1697900(Scipy), pp. 51–56 (2010)
17. Miller, C., Valasek, C.: Remote exploitation of an unaltered passenger vehicle.
Technical report, IOActive Labs, August 2015
18. uter, M., Asaj, N.: Entropy-based anomaly detection for in-vehicle networks. In:
2011 IEEE Intelligent Vehicles Symposium (IV), pp. 1110–1115, June 2011
19. Narayanan, S.N., Mittal, S., Joshi, A.: OBD SecureAlert: an anomaly detection
system for vehicles. In: IEEE Workshop on Smart Service Systems, SmartSys 2016,
May 2016
20. Oliphant, T.E.: Guide to NumPy. Methods 1, 378 (2010)
21. Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., et al.: Scikit-learn:
machine learning in Python. J. Mach. Learn. Res. 12(Oct), 2825–2830 (2012)
22. Rieke, R., Seidemann, M., Talla, E.K., Zelle, D., Seeger, B.: Behavior analysis for
safety and security in automotive systems. In: 2017 25nd Euromicro International
Conference on Parallel, Distributed and Network-Based Processing (PDP), pp.
381–385. IEEE Computer Society, March 2017
23. Song, H., Kim, H., Kim, H.: Intrusion detection system based on the analysis of
time intervals of CAN messages for in-vehicle network, pp. 63–68. IEEE Computer
Society, March 2016
24. Studnia, I., Alata, E., Nicomette, V., Kaˆaniche, M., Laarouchi, Y.: A language-
based intrusion detection approach for automotive embedded networks. In: The
21st IEEE Pacific Rim International Symposium on Dependable Computing,
PRDC 2015, November 2014
Comparative Study of Machine Learning Methods 101
25. Studnia, I., Nicomette, V., Alata, E., Deswarte, Y., Kaˆaniche, M., Laarouchi, Y.:
Security of embedded automotive networks: state of the art and a research proposal.
In: ROY, M. (ed.) SAFECOMP 2013 - Workshop CARS of the 32nd International
Conference on Computer Safety, Reliability and Security, September 2013
26. Taylor, A., Leblanc, S.P., Japkowicz, N.: Probing the limits of anomaly detectors
for automobiles with a cyber attack framework. IEEE Intell. Syst. PP(99), 1 (2018)
27. Theissler, A.: Anomaly detection in recordings from in-vehicle networks. In: Pro-
ceedings of Big Data Applications and Principles First International Workshop,
BIGDAP 2014, 11–12 September 2014, Madrid, Spain (2014)
28. Waskom, M., Meyer, K., Hobson, P., Halchenko, Y., et al.: Seaborn: v0.5.0, Novem-
ber 2014
29. Wei, Z., Yang, Y., Rehana, Y., Wu, Y., Weng, J., Deng, R.H.: IoVShield: an ecient
vehicular intrusion detection system for self-driving (short paper). In: Liu, J.K.,
Samarati, P. (eds.) ISPEC 2017. LNCS, vol. 10701, pp. 638–647. Springer, Cham
(2017). https://doi.org/10.1007/978-3-319- 72359-4 39
30. Wolf, M., Weimerskirch, A., Paar, C.: Security in automotive bus systems. In:
Proceedings of the Workshop on Embedded Security in Cars (July), pp. 1–13 (2004)
... Although the supervised learning-based IDS has a low false alarm rate (FAR), the kind methods have a weak ability to detect unknown attacks [10][11][12]. In contrast, the unsupervised learning-based IDS can identify various anomaly traffic through a deviation between normal behavior and abnormal behavior, but with a relatively high false positive rate (FPR) [13][14][15]. Moreover, many works design the attack model by injecting a high volume of bursty anomaly messages, which could cause false-high performance [16][17][18]. ...
Article
Full-text available
Intrusion Detection Systems (IDS) are considered essential security components for the Controller Area Network (CAN) in vehicular communications. However, existing methods struggle to adapt to varied attack scenarios and accurately detect low-volume attacks. In this paper, we introduce Meta-IDS, a novel IDS that employs meta-learning via the Meta-SGD algorithm to enhance adaptability across a diverse spectrum of cyber threats. Our method includes a bi-level optimization technique: the inner level optimizes detection accuracy for specific attack scenarios, while the outer level adjusts meta-parameters to ensure generalizability across different scenarios. To model low-volume attacks, we devise the Attack Prominence Score (APS), which identifies subtle attack patterns. Extensive experimental results show that the proposed method exhibits promising detection performance, with efficient tuning and rapid adaptation to different attack scenarios, especially low-volume attacks. Furthermore, Meta-IDS demonstrates competitive scalability in intelligent transportation systems, supporting larger network infrastructures and high throughput. Real-time vehicle-level evaluations also show that it is lightweight for vehicular networks.
Conference Paper
The rapid proliferation of electric vehicles (EVs) necessitates advanced charging infrastructure, which is increasingly reliant on cloud-based technologies and the Internet of Things (IoT). However, these systems are vulnerable to cyber attacks that could have severe repercussions, including power grid failures. This paper addresses the security vulnerabilities inherent in the EV charging system. We propose a novel hybrid security solution tailored for cloud-based EV charging systems that integrates time-series-based regression models with classification algorithms to detect and mitigate both power consumption anomalies and network intrusions effectively. Our approach includes a comprehensive analysis to identify vulnerabilities and threats for cloud-based EV systems, a dual-model system for anomaly and intrusion detection, and a feedback-based threshold adjustment mechanism to assist overflow and anomaly identification. We provide a detailed analysis of data communication threats from a cloud perspective, design a robust security model, and evaluate various models to select the most effective ones for real-time security management. The finally decided models present excellent accuracy and practical value. Our findings contribute to enhancing the resilience of EV charging systems against cyber-physical threats, ensuring more reliable and secure operations.
Article
The development of intrusion detection systems (IDS) for the in-vehicle Controller Area Network (CAN) bus is one of the main efforts being taken to secure the in-vehicle network against various cyberattacks, which have the potential to cause vehicles to malfunction and result in dangerous accidents. These CAN IDS are evaluated in disparate experimental conditions that vary in terms of the workload used, the features used, the metrics reported, etc., which makes direct comparison difficult. Therefore, there have been several benchmarking frameworks and comparative studies designed to evaluate CAN IDS in similar experimental conditions to understand their relative performance and facilitate the selection of the best CAN IDS for implementation in automotive networks. This work provides a comprehensive survey of CAN IDS benchmarking frameworks and comparative studies in the current literature. A CAN IDS evaluation design space is also proposed in this work, which draws from the wider CAN IDS literature. This is not only expected to serve as a guide for designing CAN IDS evaluation experiments but is also used for categorising current benchmarking efforts. The surveyed works have been discussed on the basis of the five aspects in the design space – namely, IDS type, attack model, evaluation type, workload generation, and evaluation metrics – and recommendations for future work have been identified.
Article
Full-text available
According to research, the vast majority of road accidents (90%) are the result of human error, with only a small percentage (2%) being caused by malfunctions in the vehicle. Smart vehicles have gained significant attention as potential solutions to address such issues. In the future of transportation, travel comfort and road safety will be ensured while also offering several value-added services. The automotive industry has undergone a significant transformation through the use of emerging technologies and wireless communication channels, resulting in vehicles becoming more interconnected, intelligent, and safe. However, these technologies and communication systems are susceptible to numerous security attacks. The objective of this paper is to present a comprehensive overview of the smart vehicle’s architecture, encompassing emerging technologies and security challenges and solutions associated with smart vehicles. There has been a significant surge in the utilization of machine learning techniques in smart vehicles. We categorically discuss common security measures, including machine learning and deep learning based solutions that have been mentioned in the literature and implemented against security threats on smart vehicles. This paper has also been titled a tutorial due to its layout, which begins with covering preliminary knowledge, terminologies, and encompassing technologies required to comprehend smart vehicles. Following this, the paper addresses the overall challenges associated with smart vehicles and then focuses on security issues. In terms of solutions, the paper discusses overall solutions to security issues in smart vehicles before delving into a specific solution based on machine learning and deep learning.
Article
Full-text available
A federated learning-based intrusion detection system (FL-IDS) is introduced in this paper to enhance the security of vehicular networks in the context of IoT edge device implementations. The FL-IDS system protects data privacy by using local learning, where devices share only model updates with an aggregation server. This server then generates an enhanced detection model. The FL-IDS system also incorporates machine learning (ML) and deep learning (DL) classifiers, namely logistic regression (LR) and convolutional neural networks (CNN), to prevent attacks in transportation IoT environments. The performance of the proposed IDS was evaluated using two different datasets, NSL-KDD and Car-Hacking. The model evaluation has been evaluated based on the accuracy and loss parameters. The results showthat the FL-IDS system outperforms traditional centralized machine learning and deep learning approaches regarding accuracy and privacy protection.
Article
Full-text available
The development of Intrusion Detection Systems (IDS) for in-vehicle buses has gained a lot of momentum in recent years as the number of reported vulnerabilities and the degree of interconnectivity for modern vehicles are on the rise. Since intrusion detection is resource consuming, it can be performed on computationally capable Android head units that are now present inside vehicles. Moreover, these units are connected to the internet, which enables the use of more complex algorithms that run in cloud environments. In this work we develop one such approach: an IDS that consists of a locally installed copy, running on head units, and a centralized instance of it that runs in the cloud and monitors traffic for groups of similar vehicles. Additionally, the centralized instance is part of a cloud service for intrusion detection which is continuously updated with the most recent types of attacks. The classification results of the cloud-based service are further analyzed by an incident response team which confirms the presence of known attacks, analyzes new types of attacks and assesses their impact. The output of this activity is stored on the Blockchain as ISO/SAE 21434 compliant reports, ensuring the transparency and traceability of the reported incidents.
Preprint
Full-text available
In the rapidly evolving landscape of vehicular communications, the widespread use of the Controller Area Network (CAN) in modern vehicles has revealed significant security vulnerabilities. However, existing Intrusion Detection Systems (IDS) struggle to adapt to varied attack scenarios and precisely detect low-volume attacks. In this paper, we introduce a novel IDS that employs meta-learning via the Meta-SGD algorithm, enhancing adaptability across a diverse spectrum of cyber threats, called Meta-IDS. Specifically, our methodology includes a bi-level optimization technique where the inner level focuses on optimizing detection accuracy for specific attack scenarios, and the outer level adjusts meta-parameters to ensure generalizability across different scenarios. For modeling low-volume attacks, we devise the Attack Prominence Score (APS), identifying subtle attack patterns with a threshold of APS \textgreater 7, allowing for precise differentiation of these attacks. The extensive experiment results show that the proposed method facilitates efficient tuning and rapid adaptation for different modeling paradigms in few-shot scenarios. The detection performance is exceptional, with F1-scores reaching 100% across most attack scenarios, including low-volume attacks. Also, the real-time vehicle-level evaluations demonstrate its adaptability for the vehicular networks.
Conference Paper
Full-text available
The CAN bus protocol is used in modern vehiclesto control dozens of devices. It is simple enough, but this protocol lacks security mechanisms, and that is one of the main reasons that the entire vehicle is vulnerable for different cyber-physical attacks. Therefore, the detection and counteraction against suchattacks on the CAN bus protocol is an important task. In thepaper, we present an approach for visual analysis of the СAN busprotocol traffic dumps, which is based on visualization of traffictimeline segments by using radial bar charts. Experiments forvisualization of traffic dumps are presented. The advantages ofthe proposed approach are simplicity and clarity in determining attacks. The proposed model allows displaying a large amount ofdetailed information for a long-time period at one time and onone screen.
Article
Full-text available
—The increase in connectivity and complexity of modern automotive networks presents new opportunities for potential hackers trying to take over a vehicle. To protect the automotive networks from such attacks, security mechanisms, such as firewalls or secure authentication protocols may be included. However, should an attacker succeed in bypassing such measures and gain access to the internal network, these security mechanisms become unable to report about the attacks ensuing such a breach, occurring from the internal network. To complement these preventive security mechanisms, we present a non intrusive network-based intrusion detection approach fit for vehicular networks, such as the widely used CAN. Leveraging the high predictability of embedded automotive systems, we use language theory to elaborate a set of attack signatures derived from behavioural models of the automotive calculators in order to detect a malicious sequence of messages transiting through the internal network.
Article
The proliferation of computerized functions aimed at enhancing drivers’ safety and convenience has increased the number of vehicular attack surfaces accordingly. The fundamental vulnerability is caused by the fact that the controller area network protocol, a de facto standard for in-vehicle networks, does not support message origin authentication. Several methods to resolve this problem have been suggested. However, most of them require modification of the CAN protocol and have their own vulnerabilities. In this paper, we focus on securing in-vehicle CAN networks, proposing a novel automotive intrusion detection system (so-called VoltageIDS). The system leverages the inimitable characteristics of an electrical CAN signal as a fingerprint of the electronic control units. The noteworthy contributions are that VoltageIDS does not require any modification of the current system and has been validated on actual vehicles while driving on the road. VoltageIDS is also the first automotive IDS capable of distinguishing between errors and the bus-off attack. Our experimental results on a CAN bus prototype and on real vehicles show that VoltageIDS detects intrusions in the in-vehicle CAN network. Moreover, we evaluate VoltageIDS while a vehicle is moving.
Article
The increase in connectivity and complexity of modern automotive networks presents new opportunities for potential hackers trying to take over a vehicle. To protect the automotive networks from such attacks, security mechanisms, such as firewalls or secure authentication protocols may be included. However, should an attacker succeed in bypassing such measures and gain access to the internal network, these security mechanisms become unable to report about the attacks ensuing such a breach, occurring from the internal network. To complement these preventive security mechanisms, we present a non-intrusive network-based intrusion detection approach fit for vehicular networks, such as the widely used CAN. Leveraging the high predictability of embedded automotive systems, we use language theory to elaborate a set of attack signatures derived from behavioural models of the automotive calculators in order to detect a malicious sequence of messages transiting through the internal network.
Article
Modern automobiles are controlled by computers and are increasingly connected to the outside world. This makes them vulnerable to cyber-attacks. Defending cars against cyber-attacks requires a multifaceted approach to improving security, but the last line of defence is detecting those attacks within the data traffic exchanged by the vehicles controller computers. To identify this malicious traffic, we created anomaly detectors using recurrent neural networks and multivariate Markov chains. However evaluating these detectors is difficult because there are currently few examples of attack traffic. Our solution is the creation of an attack framework that describes automotive cyber-attack characteristics. This framework enables the simulation of attacks, allowing comprehensive testing of our anomaly detectors. Our approach reveals which characteristics of attacks and network traffic create the most challenging detection scenarios, and can inform future work in automotive cyber-attack detection.
Chapter
In recent years, a lot of vehicle attacks have been reported and demonstrated by researchers and whitehat hackers indicating vehicle cyber security as an important issue particularly for self-driving cars. The reason behind this extended attack vector is the multiple external interfaces of vehicles and minimal internal security protection. Hence, it is totally possible for adversaries to take full control of connected cars. In this paper, we propose an efficient Vehicular Intrusion Detection System (IDS), named as VIDS, which consists of a lightweight domain-based detection model for ECU devices and a comprehensive crossdomain-based detection model for a gateway or domain controller. The former makes use of specification periodic features of Controller Area Network (CAN) frames, while the latter exploits stream bit value features with deep learning techniques. With the use of real vehicular normal datasets and synthesized abnormal datasets for experimenting, the experimental results indicate that the proposed VIDS can achieve better detection rate over existing IDS systems. In addition, VIDS is compatible with vehicle internal CAN network.
Article
The vehicular connectivity revolution is fueling the automotive industry's most significant transformation seen in decades. However, as modern vehicles become more connected, they also become much more vulnerable to cyber-attacks. In this paper, a fully working machine learning approach is proposed to protect connected vehicles (fleets and individuals) against such attacks. We present a system that monitors different vehicle's interfaces (Network, CAN and OS), extracts relevant information based on configurable rules and sends it to a trained generative model to detect deviations from normal behavior. Using configurable data collector, we provide a higher level of data abstraction as the model is trained based on events instead of raw data, which has a noise-filtering effect and eliminates the need to retrain the model whenever a protocol changes. We present a new approach for detecting anomalies, tailored to the temporal nature of our domain. Adapting the hybrid approach of Gutflaish et al. (2017) to the fully temporal setting, we first train a Hidden Markov Model to learn normal vehicle behavior, and then a regression model to calibrate the likelihood threshold for anomaly. Using this architecture, our method detects sophisticated and realistic anomalies, which are missed by other existing methods monitoring the CAN bus only. We also demonstrate the superiority of adaptive thresholds over static ones. Furthermore, our approach scales efficiently from monitoring individual cars to serving large fleets. We demonstrate the competitive advantage of our model via encouraging empirical results.
Technical Report
TensorFlow [1] is an interface for expressing machine learning algorithms, and an implementation for executing such algorithms. A computation expressed using TensorFlow can be executed with little or no change on a wide variety of heterogeneous systems, ranging from mobile devices such as phones and tablets up to large-scale distributed systems of hundreds of machines and thousands of computational devices such as GPU cards. The system is flexible and can be used to express a wide variety of algorithms, including training and inference algorithms for deep neural network models, and it has been used for conducting research and for deploying machine learning systems into production across more than a dozen areas of computer science and other fields, including speech recognition, computer vision, robotics, information retrieval, natural language processing, geographic information extraction, and computational drug discovery. This paper describes the TensorFlow interface and an implementation of that interface that we have built at Google. The TensorFlow API and a reference implementation were released as an open-source package under the Apache 2.0 license in November, 2015 and are available at www.tensorflow.org.