5. http://pralab.diee.unica.it
Smart Fridge Caught Sending Spam
• Jan., 2014: A fridge has been caught sending spam after a web
attack managed to compromise smart gadgets
• The fridge was one of the 100,000 compromised devices used in
the spam campaign
5 http://www.bbc.com/news/technology-25780908
6. http://pralab.diee.unica.it
New Challenges for ML/PR
• We are living exciting time for ML/PR technologies
– Our work feeds a lot of consumer technologies
for personal applications
• This opens up new big possibilities
but also new security risks
• Proliferation and sophistication
of attacks and cyberthreats
– Skilled / economically-motivated
attackers (e.g., ransomware)
• Several security systems use machine learning to detect attacks
– but … is machine learning secure enough?
6
7. http://pralab.diee.unica.it
Are we ready for this?
Can we use classical pattern recognition
and machine learning techniques under
attack?
No, we cannot. We are facing an
adversarial setting…
We should learn to find secure patterns
7
9. http://pralab.diee.unica.it
Secure Patterns in Computer Security
• Similar phenomenon in machine learning and computer security
– Obfuscation and polymorphism to hide malicious content
Spam emails Malware
Start 2007 with a bang!
Make WBFS YOUR PORTFOLIO’s
first winner of the year
...
<script type="text/javascript"
src=”http://
palwas.servehttp.com//ml.php"></
script>
...
var PGuDO0uq19+PGuDO0uq20;
EbphZcei=PVqIW5sV.replace(/jTUZZ/
g,"%"); var eWfleJqh=unescape;
Var NxfaGVHq=“pqXdQ23KZril30”;
q9124=this; var SkuyuppD=
q9124["WYd1GoGYc2uG1mYGe2YnltY".r
eplace(/[Y12WlG:]/g,
"")];SkuyuppD.write(eWfleJqh(Ebph
Zcei));
...
9
10. http://pralab.diee.unica.it
• Adaptation/evolution is crucial to survive!
Arms Race
Attackers
Evasion techniques
System designers
Design of effective countermeasures
text analysis on spam emails
visual analysis of attached images
Image-based spam
…
10
12. http://pralab.diee.unica.it
Design of Learning-based Systems
Training phase
x x x
x x
x
x
x
x
x
x
x
x x x x
x
x1
x2
...
xd
pre-processing and
feature extrac-on
training data
(with labels)
classifier learning
Linear classifiers:
assign a weight to each feature
and classify a sample based
on the sign of its score
f (x) = sign(wT
x)
+1, malicious
−1, legitimate
"
#
$
%$
start
bang
portfolio
winner
year
...
university
campus
Start 2007
with a bang!
Make WBFS
YOUR
PORTFOLIO’s
first winner
of the year
...
start
bang
portfolio
winner
year
...
university
campus
1
1
1
1
1
...
0
0
x SPAM start
bang
portfolio
winner
year
...
university
campus
+2
+1
+1
+1
+1
...
-3
-4
w
12
13. http://pralab.diee.unica.it
Design of Learning-based Systems
Test phase
x1
x2
...
xd
pre-processing and
feature extrac-on
test data
x
x
x
x
x x
x
x
x
x
x
x
x x
x x
x
classifica-on and
performance evalua-on
e.g., classifica-on accuracy
Start 2007
with a bang!
Make WBFS
YOUR
PORTFOLIO’s
first winner
of the year
...
start
bang
portfolio
winner
year
...
university
campus
1
1
1
1
1
...
0
0
start
bang
portfolio
winner
year
...
university
campus
+2
+1
+1
+1
+1
...
-3
-4
+6 > 0, SPAM
(correctly classified)
x w
Linear classifiers:
assign a weight to each feature
and classify a sample based
on the sign of its score
f (x) = sign(wT
x)
+1, malicious
−1, legitimate
"
#
$
%$
13
14. http://pralab.diee.unica.it
Can Machine Learning Be Secure?
• Problem: how to evade a linear (trained) classifier?
Start 2007
with a bang!
Make WBFS
YOUR
PORTFOLIO’s
first winner
of the year
...
start
bang
portfolio
winner
year
...
university
campus
1
1
1
1
1
...
0
0
+6 > 0, SPAM
(correctly classified)
f (x) = sign(wT
x)
x
start
bang
portfolio
winner
year
...
university
campus
+2
+1
+1
+1
+1
...
-3
-4
w
x’
St4rt 2007
with a b4ng!
Make WBFS
YOUR
PORTFOLIO’s
first winner
of the year
... campus
start
bang
portfolio
winner
year
...
university
campus
0
0
1
1
1
...
0
1
+3 -4 < 0, HAM
(misclassified email)
f (x) = sign(wT
x)
14
15. http://pralab.diee.unica.it
Can Machine Learning Be Secure?
• Underlying assumption of machine learning techniques
– Training and test data are sampled from the same distribution
• In practice
– The classifier generalizes well from known examples (+ random noise)
– … but can not cope with carefully-crafted attacks!
• It should be taught how to do that
– Explicitly taking into account adversarial data manipulation
– Adversarial machine learning
• Problem: how can we assess classifier security in a more
systematic manner?
(1) M. Barreno, B. Nelson, R. Sears, A. D. Joseph, and J. D. Tygar. Can machine
learning be secure? ASIACCS 2006
(2) B. Biggio, G. Fumera, F. Roli. Security evaluation of pattern classifiers under
attack. IEEE Trans. on Knowl. and Data Engineering, 2014 15
17. http://pralab.diee.unica.it
Adversary model
• Goal of the attack
• Knowledge of the attacked system
• Capability of manipulating data
• Attack strategy as an optimization problem
Security Evaluation of Pattern Classifiers
[B. Biggio, G. Fumera, F. Roli, IEEE Trans. KDE 2014]
Bounded adversary!
C1
C2
accuracy
Performance of more secure classifiers should degrade more gracefully under attack
0
Bound on knowledge / capability (e.g., number of modified words)
performance degradation of
text classifiers in spam filtering
against a different number of
modified words in spam emails
Security Evaluation Curve
17
18. http://pralab.diee.unica.it
Adversary’s Goal
1. Security violation [Barreno et al. ASIACCS06]
– Integrity: evade detection without
compromising system operation
– Availability: classification errors
to compromise system operation
– Privacy: gaining confidential
information about system users
2. Attack’s specificity [Barreno et al. ASIACCS06]
– Targeted/Indiscriminate: misclassification of a specific set/any sample
Integrity
Availability Privacy
Spearphishing vs phishing
18
19. http://pralab.diee.unica.it
Adversary’s Knowledge
• Perfect knowledge
– upper bound on the performance degradation under attack
TRAINING DATA
FEATURE
REPRESENTATION
LEARNING
ALGORITHM
e.g., SVM
x1
x2
...
xd
x x x
x x
x
x
x
x
x
x
x
x x x x
x
- Learning algorithm
- Parameters (e.g., feature weights)
- Feedback on decisions
19
20. http://pralab.diee.unica.it
• Attack’s influence [Barreno et al. ASIACCS06]
– Manipulation of training/test data
• Constraints on data manipulation
– maximum number of samples that can be added to the training data
• the attacker usually controls only a small fraction of the training samples
– maximum amount of modifications
• application-specific constraints
in feature space
• e.g., max. number of words that
are modified in spam emails
Adversary’s Capability
d(x, !x ) ≤ dmax
x2
x1
f(x)
x
Feasible domain
x '
20
21. http://pralab.diee.unica.it
Main Attack Scenarios
• Evasion attacks
– Goal: integrity violation, indiscriminate attack
– Knowledge: perfect / limited
– Capability: manipulating test samples
e.g., manipulation of spam emails at test time to evade detection
• Poisoning attacks
– Goal: availability violation, indiscriminate attack
– Knowledge: perfect / limited
– Capability: injecting samples into the training data
e.g., send spam with some ‘good words’ to poison the anti-spam filter,
which may subsequently misclassify legitimate emails containing such
‘good words’
21
22. http://pralab.diee.unica.it
Targeted classifier: SVM
• Maximum-margin linear classifier f (x) = sign(g(x)), g(x) = wT
x + b
min
w,b
1
2
wT
w+C max 0,1− yi f (xi )( )
i
∑
1/margin classifica-on error on training data (hinge loss)
22
23. http://pralab.diee.unica.it
• Enables learning and classification
using only dot products between samples
• Kernel functions for nonlinear classification
– e.g., RBF Kernel
Kernels and Nonlinearity
w = αi yi xi
i
∑ → g(x) = αi yi x, xi
i
∑ + b
support vectors
k(x, xi ) = exp −γ x − xi
2
( )−2−1.5−1−0.500.511.5
23
24. http://pralab.diee.unica.it
Evasion A>acks
24
1. B. Biggio, I. Corona, D. Maiorca, B. Nelson, N. Srndic, P. Laskov, G. Giacinto, and F. Roli.
Evasion attacks against machine learning at test time. ECML PKDD, 2013.
2. B. Biggio et al., Security evaluation of SVMs. SVM applications. Springer, 2014
3. F. Zhang et al., Adversarial feature selection against evasion attacks, IEEE TCYB 2016.
25. http://pralab.diee.unica.it
A Simple Example
• Problem: how to evade a linear (trained) classifier?
– We have seen this already…
• But… what if the classifier is nonlinear?
– Decision functions can be arbitrarily complicated, with no clear
relationship between features (x) and classifier parameters (w)
St4rt 2007
with a b4ng!
Make WBFS
YOUR
PORTFOLIO’s
first winner
of the year
... campus
start
bang
portfolio
winner
year
...
university
campus
0
0
1
1
1
...
0
1
start
bang
portfolio
winner
year
...
university
campus
+2
+1
+1
+1
+1
...
-3
-4
+3 -4 < 0, HAM
(misclassified email)
f (x) = sign(wT
x)
x’ w
25
26. http://pralab.diee.unica.it
Gradient-descent Evasion Attacks
• Goal: maximum-confidence evasion
• Knowledge: perfect
• Attack strategy:
• Non-linear, constrained optimization
– Gradient descent: approximate
solution for smooth functions
• Gradients of g(x) can be analytically
computed in many cases
– SVMs, Neural networks
−2−1.5−1−0.500.51
x
f (x) = sign g(x)( )=
+1, malicious
−1, legitimate
"
#
$
%$
min
x'
g(x')
s.t. d(x, x') ≤ dmax
x '
26
27. http://pralab.diee.unica.it
Computing Descent Directions
Support vector machines
Neural networks
x1
xd
δ1
δk
δm
xf g(x)
w1
wk
wm
v11
vmd
vk1
……
……
g(x) = αi yik(x,
i
∑ xi )+ b, ∇g(x) = αi yi∇k(x, xi )
i
∑
g(x) = 1+exp − wkδk (x)
k=1
m
∑
#
$
%
&
'
(
)
*
+
,
-
.
−1
∂g(x)
∂xf
= g(x) 1− g(x)( ) wkδk (x) 1−δk (x)( )vkf
k=1
m
∑
RBF kernel gradient: ∇k(x,xi
) = −2γ exp −γ || x − xi
||2
{ }(x − xi
)
27
28. http://pralab.diee.unica.it
An Example on Handwritten Digits
• Nonlinear SVM (RBF kernel) to discriminate between ‘3’ and ‘7’
• Features: gray-level pixel values
– 28 x 28 image = 784 features
Before attack (3 vs 7)
5 10 15 20 25
5
10
15
20
25
After attack, g(x)=0
5 10 15 20 25
5
10
15
20
25
After attack, last iter.
5 10 15 20 25
5
10
15
20
25
0 500
−2
−1
0
1
2
g(x)
number of iterations
Number of modified
gray-level values
Few modifications are
enough to evade detection!
… without even mimicking
the targeted class (‘7’)
28
29. http://pralab.diee.unica.it
Bounding the Adversary’s Knowledge
Limited knowledge attacks
• Only feature representation and learning algorithm are known
• Surrogate data sampled from the same distribution as the
classifier’s training data
• Classifier’s feedback to label surrogate data
PD(X,Y)data
Surrogate
training data
f(x)
Send queries
Get labels
Learn
surrogate
classifier
f’(x)
29
30. http://pralab.diee.unica.it
Experiments on PDF Malware Detection
• PDF: hierarchy of interconnected objects (keyword/value pairs)
• Adversary’s capability
– adding up to dmax objects to the PDF
– removing objects may
compromise the PDF file
(and embedded malware code)!
/Type 2
/Page 1
/Encoding 1
…
13 0 obj
<< /Kids [ 1 0 R 11 0 R ]
/Type /Page
... >> end obj
17 0 obj
<< /Type /Encoding
/Differences [ 0 /C0032 ] >>
endobj
Features: keyword count
min
x'
g(x')
s.t. d(x, x') ≤ dmax
x ≤ x'
30
31. http://pralab.diee.unica.it
Experiments on PDF Malware Detection
• Dataset: 500 malware samples (Contagio), 500 benign (Internet)
– Targeted (surrogate) classifier trained on 500 (100) samples
• Evasion rate (FN) at FP=1% vs max. number of added keywords
– Averaged on 5 repetitions
– Perfect knowledge (PK); Limited knowledge (LK)
0 10 20 30 40 50
0
0.2
0.4
0.6
0.8
1
dmax
FN
SVM (Linear), λ=0
PK (C=1)
LK (C=1)
Linear SVM
0 10 20 30 40 50
0
0.2
0.4
0.6
0.8
1
dmax
FN
SVM (RBF), λ=0
PK (C=1)
LK (C=1)
Nonlinear SVM (RBF Kernel)
Number of added keywords to each PDFNumber of added keywords to each PDF
31
32. http://pralab.diee.unica.it
Security Measures against Evasion Attacks
• Multiple Classifier Systems (MCSs)
– Feature Equalization
[Kolcz and Teo, CEAS 2009; Biggio et al., IJMLC 2010]
– 1.5-class classification
[Biggio et al., MCS 2015]
• Adversarial Feature Selection
[Zhang et al., IEEE TCYB 2016]
• Learning with Invariances: Nightmare at Test Time (InvarSVM)
– Robust optimization (zero-sum games)
[Globerson and Teo, ICML 2006]
• Game Theory (NashSVM)
– Classifier vs. Adversary (non-zero-sum games)
[Brueckner et al., JMLR 2012]
32
33. http://pralab.diee.unica.it
MCSs for Feature Equalization
• Rationale: more uniform feature weight distributions require the
attacker to modify more features to evade detection
33
1. Kolcz and C. H. Teo. Feature weighting for improved classifier robustness, CEAS 2009.
2. B. Biggio, G. Fumera, and F. Roli. Multiple classifier systems for robust classifier
design in adversarial environments. Int’l J. Mach. Learn. Cyb., 1(1):27–41, 2010.
buy
cheap
kid
game
buy cheap
kid game
Buy cheap!
…
weights
weights
1
K
fk (x)
k=1
K
∑
f1(x) = wi
1
xi + w0
1
∑
fK (x) = wi
K
xi + w0
K
∑
…
DATA
bagging,
RSMMCSs can be
exploited to
obtain more
uniform feature
weights
34. http://pralab.diee.unica.it
Wrapper-based Adversarial Feature Selection (WAFS)
[F. Zhang et al., IEEE TCYB ‘16]
• Rationale: feature selection based on accuracy and security
– wrapper-based backward/forward feature selection
– main limitation: computational complexity
34
0 5 10 15 20
0
0.2
0.4
0.6
0.8
1
Feature set size: 100
TPatFP=1%
Traditional (PK)
WAFS (PK)
0 5 10 15 20
0
0.2
0.4
0.6
0.8
1
TPatFP=1%
max. num. of modified words
Traditional (LK)
WAFS (LK)
0 5 10 15 20
0
0.2
0.4
0.6
0.8
1
Feature set size: 200
TPatFP=1%
Traditional (PK)
WAFS (PK)
0 5 10 15 20
0
0.2
0.4
0.6
0.8
1
TPatFP=1%
max. num. of modified words
Traditional (LK)
WAFS (LK)
0.
0.
0.
0.
TPatFP=1%
0.
0.
0.
0.
TPatFP=1%
20
)
0 5 10 15 20
0
0.2
0.4
0.6
0.8
1
Feature set size: 200
TPatFP=1%
Traditional (PK)
WAFS (PK)
0.6
0.8
1
P=1%
Traditional (LK)
WAFS (LK)
0 5 10 15 20
0
0.2
0.4
0.6
0.8
1
Feature set size: 300
TPatFP=1%
Traditional (PK)
WAFS (PK
0.6
0.8
1
P=1%
Traditional (LK)
WAFS (LK)
0
0.2
0.4
0.6
0.8
1
TPatFP=1%
0.6
0.8
1
P=1%
Experimental results on spam filtering (linear SVM)
35. http://pralab.diee.unica.it
1.5-class Classification
Underlying rationale
35
2−class classification
−5 0 5
−5
0
5
1−class classification (legitimate)
−5 0 5
−5
0
5
• 2-class classification is usually more accurate in the absence of attack
• … but potentially more vulnerable under attack (not enclosing legitimate data)
1.5C classification (MCS)
−5 0 5
−5
0
5
1.5-class classification aims at retaining high accuracy and security under attack
36. http://pralab.diee.unica.it
data
1C Classifier
(malicious)
Feature
Extraction
malicious
1C Classifier
(legitimate)
2C Classifier
1C Classifier
(legitimate)
legitimate
x
g1(x)
g2(x)
g3(x)
g(x) ≥ t
g(x)
true
false
Secure 1.5-class Classification with MCSs
• Heuristic approach to 1.5-class classification
36
0 5 10 15 20 25 30
0
0.2
0.4
0.6
0.8
1
maximum number of modified words
AUC1%
(PK)
2C SVM
1C SVM (L)
1C SVM (M)
1.5C MCS
0 5 10 15 20 25 30
0
0.2
0.4
0.6
0.8
1
maximum number of modified words
AUC
1%
(LK)
2C SVM
1C SVM (L)
1C SVM (M)
1.5C MCS
Spam filtering
37. http://pralab.diee.unica.it
Poisoning Machine Learning
37
1. B. Biggio, B. Nelson, P. Laskov. Poisoning attacks against SVMs. ICML, 2012
2. B. Biggio et al., Security evaluation of SVMs. SVM applications. Springer, 2014
3. H. Xiao et al., Is feature selection secure against training data poisoning? ICML, 2015
38. http://pralab.diee.unica.it
Classifier
Web Server
Tr
HTTP requests
Poisoning Attacks against SVMs
[B. Biggio, B. Nelson, P. Laskov, ICML 2012]
h#p://www.vulnerablehotel.com/components/
com_hbssearch/longDesc.php?h_id=1&
id=-2%20union%20select%20concat%28username,
0x3a,password%29%20from%20jos_users--
malicious
h#p://www.vulnerablehotel.com/login/
legitimate
✔
✔
38
39. http://pralab.diee.unica.it
• Poisoning classifier to cause denial of service
Classifier
Web Server
Tr
HTTP requests
poisoning a#ack
Poisoning Attacks against SVMs
[B. Biggio, B. Nelson, P. Laskov, ICML 2012]
h#p://www.vulnerablehotel.com/components/
com_hbssearch/longDesc.php?h_id=1&
id=-2%20union%20select%20concat%28username,
0x3a,password%29%20from%20jos_users--
legitimate
h#p://www.vulnerablehotel.com/login/
malicious
✖
✖
h#p://www.vulnerablehotel.com/login/components/
h#p://www.vulnerablehotel.com/login/longDesc.php?h_id=1&
39
40. http://pralab.diee.unica.it
• Adversary model
– Goal: to maximize classification error (availiability, indiscriminate)
– Knowledge: perfect knowledge (trained SVM and TR set are known)
– Capability: injecting samples into TR
• Attack strategy
– optimal attack point xc in TR that maximizes classification error
xc
classifica-on error = 0.039 classifica-on error = 0.022
Adversary Model and Attack Strategy
40
41. http://pralab.diee.unica.it
• Adversary model
– Goal: to maximize classification error (availiability, indiscriminate)
– Knowledge: perfect knowledge (trained SVM and TR set are known)
– Capability: injecting samples into TR
• Attack strategy
– optimal attack point xc in TR that maximizes classification error
classifica-on error = 0.022
xc
classifica-on error as a func-on of xc
Adversary Model and Attack Strategy
41
42. http://pralab.diee.unica.it
• Max. classification error L(xc)
w.r.t. xc through gradient ascent
• Gradient is not easy to compute
– The training point affects the
classification function
– Details of the derivation
are in the paper
Poisoning Attack Algorithm
xc
(0)
xc
xc
(0) xc
42 1. B. Biggio, B. Nelson, P. Laskov. Poisoning attacks against SVMs. ICML, 2012
43. http://pralab.diee.unica.it
Experiments on the MNIST digits
Single-point attack
• Linear SVM; 784 features; TR: 100; VAL: 500; TS: about 2000
– ‘0’ is the malicious (attacking) class
– ‘4’ is the legitimate (attacked) one
xc
(0)
xc
43
44. http://pralab.diee.unica.it
Experiments on MNIST digits
Multiple-point attack
• Linear SVM; 784 features; TR: 100; VAL: 500; TS: about 2000
– ‘0’ is the malicious (attacking) class
– ‘4’ is the legitimate (attacked) one
44
45. http://pralab.diee.unica.it
Poisoning linear models for feature selection
[H. Xiao et al., ICML ’15]
• Linear models
– Select features according to |w|
45
LASSO
Tibshirani, 1996
Ridge Regression
Hoerl & Kennard, 1970
Elas<c Net
Zou & Hastie, 2005
A#acker maximizes
classifica-on error
w.r.t. the a#ack point
xc (gradient ascent)
46. http://pralab.diee.unica.it
Experiments on PDF Malware Detection
• PDF: hierarchy of interconnected objects (keyword/value pairs)
• Learner’s task: to classify benign vs malware PDF files
• Attacker’s task: to maximize classification error by injecting
poisoning attack samples
– Only feature increments are considered (object insertion)
• Object removal may compromise the PDF file
/Type 2
/Page 1
/Encoding 1
…
13 0 obj
<< /Kids [ 1 0 R 11 0 R ]
/Type /Page
... >> end obj
17 0 obj
<< /Type /Encoding
/Differences [ 0 /C0032 ] >>
endobj
Features: keyword counts
46
Maiorca et al., 2012; 2013;
Smutz & Stavrou, 2012;
Srndic & Laskov, 2013
48. http://pralab.diee.unica.it
Security Measures against Poisoning
• Rationale: poisoning injects outlying training samples
• Two main strategies for countering this threat
1. Data sanitization: remove poisoning samples from training data
• Bagging for fighting poisoning attacks
• Reject-On-Negative-Impact (RONI) defense
2. Robust Learning: learning algorithms that are robust in the presence
of poisoning samples
48
xc
(0)
xc
xc
(0) xc
49. http://pralab.diee.unica.it
Security Measures against Poisoning
Data Sanitization :: Multiple Classifier Systems
• (Weighted) Bagging for fighting poisoning attacks
– Underlying idea: resampling outlying samples with lower probability
• Two-step algorithm:
1. Density estimation to assign lower resampling weights to outliers
2. Bagging to train an MCS
• Promising results on
spam (see plot) and
web-based intrusion detection
– S: standard classifier
– B: standard bagging
– WB: weighted bagging
(numbers in the legend
correspond to different ensemble sizes)
49
1. B. Biggio, I. Corona, G. Fumera, G. Giacinto, and F. Roli. Bagging classifiers for
fighting poisoning attacks in adversarial classification tasks. MCS, 2011.
50. http://pralab.diee.unica.it
A>acking Clustering
50
1. B. Biggio, I. Pillai, S. R. Bulò, D. Ariu, M. Pelillo, and F. Roli. Is data clustering in
adversarial settings secure? AISec, 2013
2. B. Biggio, S. R. Bulò, I. Pillai, M. Mura, E. Z. Mequanint, M. Pelillo, and F. Roli.
Poisoning complete-linkage hierarchical clustering. S+SSPR, 2014
51. http://pralab.diee.unica.it
Attacking Clustering
• So far, we have considered supervised learning
– Training data consisting of samples and class labels
• In many applications, labels are not available or costly to obtain
– Unsupervised learning
• Training data only include samples – no labels!
• Malware clustering
– To identify variants of existing malware or new malware families
x x x
x x
x
x
x
x
x
x
x
x x x x
x
x1
x2
...
xd
feature extraction
(e.g., URL length,
num. of parameters, etc.)
data collection
(honeypots)
clustering of
malware families
(e.g., similar HTTP
requests)
data analysis /
countermeasure design
(e.g., signature generation)
for each cluster
if …
then …
else …
51
52. http://pralab.diee.unica.it
Is Data Clustering Secure?
• Attackers can poison input data to subvert malware clustering
x
x
x
x x
x
x
x
x
x
x
x x
x
x x
x
x1
x2
...
xd
feature extraction
(e.g., URL length,
num. of parameters, etc.)
data collection
(honeypots)
clustering of
malware families
(e.g., similar HTTP
requests)
data analysis /
countermeasure design
(e.g., signature generation)
for each cluster
if …
then …
else …
Well-cra2ed HTTP requests to subvert clustering
h#p://www.vulnerablehotel.com/…
h#p://www.vulnerablehotel.com/…
h#p://www.vulnerablehotel.com/…
h#p://www.vulnerablehotel.com/…
… is significantly
compromised
… becomes
useless (too many
false alarms, low
detection rate)
52
53. http://pralab.diee.unica.it
Our Work
• A framework to identify/design attacks against clustering algorithms
– Poisoning: add samples to maximally compromise the clustering output
– Obfuscation: hide samples within existing clusters
• Some clustering algorithms can be very sensitive to poisoning!
– single- and complete-linkage hierarchical clustering can be easily
compromised by creating heterogeneous clusters
• Details on the attack derivation and implementation are in the papers
Clustering on untainted data (80 samples) Clustering after adding 10 attack samples
53
1. B. Biggio et al. Is data clustering in adversarial settings secure? AISec, 2013
2. B. Biggio et al. Poisoning complete-linkage hierarchical clustering. S+SSPR, 2014
54. http://pralab.diee.unica.it
Conclusions and Future Work
• Learning-based systems are vulnerable to well-crafted,
sophisticated attacks devised by skilled attackers
– … that exploit specific vulnerabilities of machine learning algorithms!
Secure learning algorithms
Attacks against learning
54