SlideShare une entreprise Scribd logo
1  sur  40
Télécharger pour lire hors ligne
Pa#ern 
Recogni-on 
and 
Applica-ons 
Lab 
On Learning and Recognition of Secure Patterns 
University 
of 
Cagliari, 
Italy 
Department 
of 
Electrical 
and 
Electronic 
Engineering 
BaAsta 
Biggio 
Dept. 
Of 
Electrical 
and 
Electronic 
Engineering 
University 
of 
Cagliari, 
Italy 
Sco#sdale, 
Arizona, 
AISec 
2014 
US, 
Nov., 
7 
2014
Secure Patterns in Nature 
• Learning of secure patterns is a well-known problem in nature 
– Mimicry and camouflage 
– Arms race between predators and preys 
http://pralab.diee.unica.it 
2
Secure Patterns in Computer Security 
• Similar phenomenon in machine learning and computer security 
– Obfuscation and polymorphism to hide malicious content 
Spam emails Malware 
Start 2007 with a bang! 
Make WBFS YOUR PORTFOLIO’s 
first winner of the year 
..." 
http://pralab.diee.unica.it 
<script type="text/javascript" 
src=”http:// 
palwas.servehttp.com//ml.php"></ 
script>" 
... 
var PGuDO0uq19+PGuDO0uq20; 
EbphZcei=PVqIW5sV.replace(/jTUZZ/ 
g,"%"); var eWfleJqh=unescape;" 
Var NxfaGVHq=“pqXdQ23KZril30”;" 
q9124=this; var SkuyuppD= 
q9124["WYd1GoGYc2uG1mYGe2YnltY".r 
eplace(/[Y12WlG:]/g, 
"")];SkuyuppD.write(eWfleJqh(Ebph 
Zcei));" 
..." 
3
Arms Race 
• Adaptation/evolution is crucial to survive! 
http://pralab.diee.unica.it 
Attackers 
Evasion techniques 
System designers 
Design of effective countermeasures 
text analysis on spam emails 
visual analysis of attached images 
Image-based spam 
… 
4
Machine 
Learning 
in 
Computer 
Security 
http://pralab.diee.unica.it 
5
Design of Learning-based Systems 
Training phase 
http://pralab.diee.unica.it 
xx 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x1 
x2 
... 
xd 
pre-­‐processing 
and 
feature 
extrac-on 
training 
data 
(with 
labels) 
classifier 
learning 
Linear 
classifiers: 
assign 
a 
weight 
to 
each 
feature 
and 
classify 
a 
sample 
based 
on 
the 
sign 
of 
its 
score 
f (x) = sign(wT x) 
+1, malicious 
−1, legitimate 
"# $ 
%$ 
start" 
bang 
portfolio 
winner" 
year 
... 
university 
campus" 
Start 2007 
with a bang! 
Make WBFS 
YOUR 
PORTFOLIO’s 
first winner 
of the year 
..." 
start" 
bang 
portfolio 
winner" 
year 
... 
university 
campus" 
1 
1" 1" 
1" 
1" 
..." 
0" 
0" 
SPAM 
x 
start" 
bang 
portfolio 
winner" 
year 
... 
university 
campus" 
w 
+2 
+1" +1" 
+1" 
+1" 
..." 
-3" 
-4" 
6
Design of Learning-based Systems 
Test phase 
http://pralab.diee.unica.it 
x1 
x2 
... 
xd 
pre-­‐processing 
and 
feature 
extrac-on 
test 
data 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
classifica-on 
and 
performance 
evalua-on 
e.g., 
classifica-on 
accuracy 
Start 2007 
with a bang! 
Make WBFS 
YOUR 
PORTFOLIO’s 
first winner 
of the year 
..." 
start" 
bang 
portfolio 
winner" 
year 
... 
university 
campus" 
1 
1" 1" 
1" 
1" 
..." 
0" 
0" 
start" 
bang 
portfolio 
winner" 
year 
... 
university 
campus" 
+2 
+1" +1" 
+1" 
+1" 
..." 
-3" 
-4" 
+6 > 0, SPAM" 
(correctly 
classified) 
x 
w 
Linear 
classifiers: 
assign 
a 
weight 
to 
each 
feature 
and 
classify 
a 
sample 
based 
on 
the 
sign 
of 
its 
score 
f (x) = sign(wT x) 
+1, malicious 
−1, legitimate 
"# $ 
%$ 
7
Can Machine Learning Be Secure? 
• Problem: how to evade a linear (trained) classifier? 
Start 2007 
with a bang! 
Make WBFS 
YOUR 
PORTFOLIO’s 
first winner 
of the year 
..." 
http://pralab.diee.unica.it 
start" 
bang 
portfolio 
winner" 
year 
... 
university 
campus" 
1 
1" 1" 
1" 
1" 
..." 
0" 
0" 
f (x) = sign(wT x) 
+6 > 0, SPAM" 
(correctly 
classified) 
x 
start" 
bang 
portfolio 
winner" 
year 
... 
university 
campus" 
w 
+2 
+1" +1" 
+1" 
+1" 
..." 
-3" 
-4" 
x’ 
St4rt 2007 
with a b4ng! 
Make WBFS 
YOUR 
PORTFOLIO’s 
first winner 
of the year 
... campus! 
start" 
bang 
portfolio 
winner" 
year 
... 
university 
campus" 
0 
0! 1" 
1" 
1" 
..." 
0" 
1! 
f (x) = sign(wT x) 
+3 -4 < 0, HAM" 
(misclassified 
email) 
8
Can Machine Learning Be Secure? 
• Underlying assumption of machine learning techniques 
– Training and test data are sampled from the same distribution 
• In practice 
– The classifier generalizes well from known examples (+ random noise) 
– … but can not cope with carefully-crafted attacks! 
• It should be taught how to do that 
– Explicitly taking into account adversarial data manipulation 
– Adversarial machine learning 
• Problem: how can we assess classifier security in a more 
systematic manner? 
http://pralab.diee.unica.it 
(1) M. Barreno, B. Nelson, R. Sears, A. D. Joseph, and J. D. Tygar. Can machine 
learning be secure? ASIACCS 2006 
(2) B. Biggio, G. Fumera, F. Roli. Security evaluation of pattern classifiers under 
attack. IEEE Trans. on Knowl. and Data Engineering, 2014 9
Security 
Evalua>on 
of 
Pa@ern 
Classifiers 
http://pralab.diee.unica.it 
(1) B. Biggio, G. Fumera, F. Roli. Security evaluation of pattern classifiers under 
attack. IEEE Trans. on Knowl. and Data Engineering, 2014 
(2) B. Biggio et al., Security evaluation of SVMs. SVM applications. Springer, 2014 
10
Security Evaluation of Pattern Classifiers 
[B. Biggio, G. Fumera, F. Roli, IEEE Trans. KDE 2014] 
Adversary model 
• Goal of the attack 
• Knowledge of the attacked system 
• Capability of manipulating data 
• Attack strategy as an optimization problem 
Performance of more secure classifiers should degrade more gracefully under attack 
C2 
http://pralab.diee.unica.it 
Bounded adversary! 
C1 
accuracy 
0 
performance degradation of 
text classifiers in spam filtering 
against a different number of 
modified words in spam emails 
Security Evaluation Curve 
Bound on knowledge / capability (e.g., number of modified words) 
11
Adversary’s Goal 
1. Security violation [Barreno et al. ASIACCS06] 
– Integrity: evade detection without 
compromising system operation 
– Availability: classification errors 
to compromise system operation 
– Privacy: gaining confidential 
information about system users 
2. Attack’s specificity [Barreno et al. ASIACCS06] 
– Targeted/Indiscriminate: misclassification of a specific set/any sample 
http://pralab.diee.unica.it 
Integrity 
Availability Privacy 
Spearphishing 
vs 
phishing 
12
Adversary’s Knowledge 
TRAINING DATA 
• Perfect knowledge 
- Learning algorithm 
- Parameters (e.g. feature weights) 
- Feedback on decisions 
x 
x 
– upper bound on the performance degradation under attack 
http://pralab.diee.unica.it 
FEATURE 
REPRESENTATION 
LEARNING 
ALGORITHM 
e.g., SVM 
x1 
x2 
... 
xd 
xx 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
13
Adversary’s Capability 
• Attack’s influence [Barreno et al. ASIACCS06] 
– Manipulation of training/test data 
• Constraints on data manipulation 
– maximum number of samples that can be added to the training data 
• the attacker usually controls only a small fraction of the training samples 
– maximum amount of modifications 
• application-specific constraints 
in feature space 
• e.g., max. number of words that 
are modified in spam emails 
http://pralab.diee.unica.it 
x ' 
d(x, x!) ≤ dmax 
x2 
x1 
f(x) 
x 
Feasible domain 
14
Main Attack Scenarios 
• Evasion attacks 
– Goal: integrity violation, indiscriminate attack 
– Knowledge: perfect / limited 
– Capability: manipulating test samples 
e.g., manipulation of spam emails at test time to evade detection 
• Poisoning attacks 
– Goal: availability violation, indiscriminate attack 
– Knowledge: perfect / limited 
– Capability: injecting samples into the training data 
e.g., send spam with some ‘good words’ to poison the anti-spam filter, 
which may subsequently misclassify legitimate emails containing such 
‘good words’ 
http://pralab.diee.unica.it 
15
Targeted classifier: SVM 
• Maximum-margin linear classifier f (x) = sign(g(x)), g(x) = wT x + b 
min 
w,b 
1 
2 
http://pralab.diee.unica.it 
Σ 
wTw+C max(0,1− yi f (xi )) 
i 
1/margin 
classifica-on 
error 
on 
training 
data 
(hinge 
loss) 
16
Kernels and Nonlinearity 
Σ →g(x) = αi 
w = αi 
Σ + b 
• Enables learning and classification 
using only dot products between samples 
• Kernel functions for nonlinear classification 
– e.g., RBF −2 Kernel 
( 2 ) −1.5 −1 −0.5 0 0.5 1 1.5 
http://pralab.diee.unica.it 
yixi 
i 
yi x, xi 
i 
support vectors 
k(x, xi ) = exp −γ x − xi 
17
Evasion 
A@acks 
http://pralab.diee.unica.it 
18 
1. B. Biggio, I. Corona, D. Maiorca, B. Nelson, N. Srndic, P. Laskov, G. Giacinto, and 
F. Roli. Evasion attacks against machine learning at test time. ECML PKDD, 2013. 
2. B. Biggio et al., Security evaluation of SVMs. SVM applications. Springer, 2014
A Simple Example 
• Problem: how to evade a linear (trained) classifier? 
– We have seen this already… 
St4rt 2007 
with a b4ng! 
Make WBFS 
YOUR 
PORTFOLIO’s 
first winner 
of the year 
... campus! 
• But… what if the classifier is nonlinear? 
f (x) = sign(wT x) 
– Decision functions can be arbitrarily complicated, with no clear 
relationship between features (x) and classifier parameters (w) 
http://pralab.diee.unica.it 
start" 
bang 
portfolio 
winner" 
year 
... 
university 
campus" 
0 
0! 1" 
1" 
1" 
..." 
0" 
1! 
start" 
bang 
portfolio 
winner" 
year 
... 
university 
campus" 
+2 
+1" +1" 
+1" 
+1" 
..." 
-3" 
-4" 
+3 -4 < 0, HAM" 
(misclassified 
email) 
x’ 
w 
19
Gradient-descent Evasion Attacks 
• Goal: maximum-confidence evasion 
• Knowledge: perfect 
• Attack strategy: 
• Non-linear, constrained optimization 
– Gradient descent: approximate 
solution for smooth functions 
• Gradients of g(x) can be analytically 
computed in many cases 
– SVMs, Neural networks 
http://pralab.diee.unica.it 
−2 −1.5 −1 −0.5 0 0.5 1 x 
f (x) = sign(g(x)) = 
+1, malicious 
−1, legitimate 
"# $ 
%$ 
min 
x' 
g(x') 
s.t. d(x, x') ≤ dmax 
x ' 
20
Computing Descent Directions 
Support vector machines 
Σ xi )+ b, ∇g(x) = αi 
g(x) = αi 
RBF kernel gradient: ∇k(x, xi ) = −2γ exp −γ || x − xi ||2 { }(x − xi ) 
Neural networks 
x1 
xd 
http://pralab.diee.unica.it 
δ1 
δk 
δm 
xf 
g(x) 
w1 
wk 
wm 
v11 
vmd 
vk1 
… … 
… … 
yik(x, 
i 
yi∇k(x, xi ) 
Σ 
i 
mΣ 
g(x) = 1+ exp − wkδk (x) 
k=1 
# 
$ % 
& 
' ( 
) 
* + 
−1 
, 
- . 
∂g(x) 
∂xf 
mΣ 
= g(x)(1− g(x)) wkδk (x) 1−δk ( (x))vkf 
k=1 
21
An Example on Handwritten Digits 
• Nonlinear SVM (RBF kernel) to discriminate between ‘3’ and ‘7’ 
• Features: gray-level pixel values 
– 28 x 28 image = 784 features 
Before attack (3 vs 7) 
5 10 15 20 25 
5 
10 
15 
20 
25 
http://pralab.diee.unica.it 
After attack, g(x)=0 
5 10 15 20 25 
5 
10 
15 
20 
25 
After attack, last iter. 
5 10 15 20 25 
5 
10 
15 
20 
25 
0 500 
2 
1 
0 
−1 
−2 
g(x) 
nNuummbbeerr ooff imteoradtiifoiends 
gray-level values 
Few modifications are 
enough to evade detection! 
… without even mimicking 
the targeted class (‘7’) 
22
Bounding the Adversary’s Knowledge 
Limited knowledge attacks 
• Only feature representation and learning algorithm are known 
• Surrogate data sampled from the same distribution as the 
classifier’s training data 
• Classifier’s feedback to label surrogate data 
PDd(aXta,Y 
) 
http://pralab.diee.unica.it 
Surrogate 
training data 
f(x) 
Send queries 
Get labels 
Learn 
surrogate 
classifier 
f’(x) 
23
Experiments on PDF Malware Detection 
• PDF: hierarchy of interconnected objects (keyword/value pairs) 
• Adversary’s capability 
– adding up to dmax objects to the PDF 
– removing objects may 
compromise the PDF file 
(and embedded malware code)! 
http://pralab.diee.unica.it 
/Type 
2 
/Page 
1 
/Encoding 
1 
… 
13 
0 
obj 
<< 
/Kids 
[ 
1 
0 
R 
11 
0 
R 
] 
/Type 
/Page 
... 
>> 
end 
obj 
17 
0 
obj 
<< 
/Type 
/Encoding 
/Differences 
[ 
0 
/C0032 
] 
>> 
endobj 
Features: 
keyword 
count 
min 
x' 
g(x') 
s.t. d(x, x') ≤ dmax 
x ≤ x' 
24
Experiments on PDF Malware Detection 
• Dataset: 500 malware samples (Contagio), 500 benign (Internet) 
– Targeted (surrogate) classifier trained on 500 (100) samples 
• Evasion rate (FN) at FP=1% vs max. number of added keywords 
– Averaged on 5 repetitions 
– Perfect knowledge (PK); Limited knowledge (LK) 
1 
0.8 
0.6 
0.4 
0.2 
0 
0 10 20 30 40 50 
http://pralab.diee.unica.it 
d 
max 
FN 
SVM (Linear), λ=0 
PK (C=1) 
LK (C=1) 
Linear SVM 
1 
0.8 
0.6 
0.4 
0.2 
0 
Nonlinear SVM (RBF Kernel) 
0 10 20 30 40 50 
d 
max 
FN 
SVM (RBF), λ=0 
PK (C=1) 
LK (C=1) 
Number Number of added keywords to each PDF of added keywords to each PDF 
25
Poisoning 
A@acks 
against 
SVMs 
http://pralab.diee.unica.it 
26 
1. B. Biggio, B. Nelson, P. Laskov. Poisoning attacks against SVMs. ICML, 2012 
2. B. Biggio et al., Security evaluation of SVMs. SVM applications. Springer, 2014
Poisoning Attacks against SVMs 
[B. Biggio, B. Nelson, P. Laskov, ICML 2012] 
malicious 
h#p://www.vulnerablehotel.com/components/ 
com_hbssearch/longDesc.php?h_id=1& 
id=-­‐2%20union%20select%20concat%28username, 
0x3a,password%29%20from%20jos_users-­‐-­‐ 
http://pralab.diee.unica.it 
Classifier 
Web 
Server 
Tr 
HTTP 
requests 
h#p://www.vulnerablehotel.com/login/ 
legitimate 
✔ 
✔ 
27
Poisoning Attacks against SVMs 
[B. Biggio, B. Nelson, P. Laskov, ICML 2012] 
• Poisoning classifier to cause denial of service 
✖ 
malicious 
legitimate 
h#p://www.vulnerablehotel.com/login/ 
h#p://www.vulnerablehotel.com/components/ 
com_hbssearch/longDesc.php?h_id=1& 
id=-­‐2%20union%20select%20concat%28username, 
0x3a,password%29%20from%20jos_users-­‐-­‐ 
http://pralab.diee.unica.it 
Classifier 
Web 
Server 
Tr 
HTTP 
requests 
poisoning 
a#ack 
✖ 
h#p://www.vulnerablehotel.com/login/components/ 
h#p://www.vulnerablehotel.com/login/longDesc.php?h_id=1& 
28
Adversary Model and Attack Strategy 
• Adversary model 
– Goal: to maximize classification error (availiability, indiscriminate) 
– Knowledge: perfect knowledge (trained SVM and TR set are known) 
– Capability: injecting samples into TR 
• Attack strategy 
– optimal attack point xc in TR that maximizes classification error 
http://pralab.diee.unica.it 
xc 
classifica-on 
error 
= 
0.022 
classifica- on 
error 
= 
0.039 
29
Adversary Model and Attack Strategy 
• Adversary model 
– Goal: to maximize classification error (availiability, indiscriminate) 
– Knowledge: perfect knowledge (trained SVM and TR set are known) 
– Capability: injecting samples into TR 
• Attack strategy 
– optimal attack point xc in TR that maximizes classification error 
classifica-on 
error 
= 
0.022 
http://pralab.diee.unica.it 
classifica-on 
error 
as 
a 
func-on 
of 
xc 
xc 
30
Poisoning Attack Algorithm 
• Max. classification error L(xc) 
w.r.t. xc through gradient ascent 
• Gradient is not easy to compute 
– The training point affects the 
classification function 
– Details of the derivation 
are in the paper 
http://pralab.diee.unica.it 
(0) 
xc 
xc 
(0) xc 
xc 
1. B. Biggio, B. Nelson, P. Laskov. Poisoning attacks against SVMs. ICML, 2012 31
Experiments on the MNIST digits 
Single-point attack 
• Linear SVM; 784 features; TR: 100; VAL: 500; TS: about 2000 
– ‘0’ is the malicious (attacking) class 
– ‘4’ is the legitimate (attacked) one 
xc 
http://pralab.diee.unica.it 
(0) xc 
32
Experiments on the MNIST digits 
Multiple-point attack 
• Linear SVM; 784 features; TR: 100; VAL: 500; TS: about 2000 
– ‘0’ is the malicious (attacking) class 
– ‘4’ is the legitimate (attacked) one 
http://pralab.diee.unica.it 
33
http://pralab.diee.unica.it 
A@acking 
Clustering 
34 
1. B. Biggio, I. Pillai, S. R. Bulò, D. Ariu, M. Pelillo, and F. Roli. Is data clustering in 
adversarial settings secure? AISec, 2013 
2. B. Biggio, S. R. Bulò, I. Pillai, M. Mura, E. Z. Mequanint, M. Pelillo, and F. Roli. 
Poisoning complete-linkage hierarchical clustering. S+SSPR, 2014
Attacking Clustering 
• So far, we have considered supervised learning 
– Training data consisting of samples and class labels 
• In many applications, labels are not available or costly to obtain 
– Unsupervised learning 
• Training data only include samples – no labels! 
• Malware clustering 
– To identify variants of existing malware or new malware families 
http://pralab.diee.unica.it 
xx 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x1 
x2 
... 
xd 
feature extraction 
(e.g., URL length, 
num. of parameters, etc.) 
data collection 
(honeypots) 
clustering of 
malware families 
(e.g., similar HTTP 
requests) 
for 
each 
cluster 
if 
… 
then 
… 
else 
… 
data analysis / 
countermeasure design 
(e.g., signature generation) 
35
Is Data Clustering Secure? 
• Attackers can poison input data to subvert malware clustering 
http://pralab.diee.unica.it 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x 
x1 
x2 
... 
xd 
feature extraction 
(e.g., URL length, 
num. of parameters, etc.) 
data collection 
(honeypots) 
clustering of 
malware families 
(e.g., similar HTTP 
requests) 
for 
each 
cluster 
if 
… 
then 
… 
else 
… 
data analysis / 
countermeasure design 
(e.g., signature generation) 
Well-­‐cra1ed 
HTTP 
requests 
to 
subvert 
clustering 
h#p://www.vulnerablehotel.com/… 
h#p://www.vulnerablehotel.com/… 
h#p://www.vulnerablehotel.com/… 
h#p://www.vulnerablehotel.com/… 
… is significantly 
compromised 
… becomes 
useless (too many 
false alarms, low 
detection rate) 
36
Our Work 
• A framework to identify/design attacks against clustering algorithms 
– Poisoning: add samples to maximally compromise the clustering output 
– Obfuscation: hide samples within existing clusters 
• Some clustering algorithms can be very sensitive to poisoning! 
– single- and complete-linkage hierarchical clustering can be easily 
compromised by creating heterogeneous clusters 
• Details on the attack derivation and implementation are in the papers 
Clustering on untainted data (80 samples) Clustering after adding 10 attack samples 
http://pralab.diee.unica.it 
37 
1. B. Biggio et al. Is data clustering in adversarial settings secure? AISec, 2013 
2. B. Biggio et al. Poisoning complete-linkage hierarchical clustering. S+SSPR, 2014
Conclusions and Future Work 
• Learning-based systems can be vulnerable to well-crafted, 
sophisticated attacks devised by skilled attackers 
– … that exploit specific vulnerabilities of machine learning algorithms! 
• Future (and ongoing) work 
– Privacy Attacks 
– Secure Learning, Clustering and Feature Selection/Reduction 
http://pralab.diee.unica.it 
Secure learning algorithms 
Attacks against learning 
38
Joint work with … 
http://pralab.diee.unica.it 
? 
Any questions Thanks 
for 
your 
a#en-on! 
39 
… and many others
References 
1. B. Biggio, G. Fumera, and F. Roli. Security evaluation of pattern classifiers under attack. IEEE Trans. on 
Knowl. and Data Eng., 26(4):984–996, April 2014. 
2. M. Barreno, B. Nelson, R. Sears, A. D. Joseph, and J. D. Tygar. Can machine learning be secure? In 
Proc. ACM Symp. Information, Computer and Comm. Sec., ASIACCS ’06, pages 16–25, New York, 
NY, USA, 2006. ACM. 
3. M. Barreno, B. Nelson, A. Joseph, and J. Tygar. The security of machine learning. Machine Learning, 
81:121–148, 2010. 
4. B. Biggio, I. Corona, B. Nelson, B. Rubinstein, D. Maiorca, G. Fumera, G. Giacinto, and F. Roli. Security 
evaluation of support vector machines in adversarial environments. In Y. Ma and G. Guo, eds, 
Support Vector Machines Applications, pages 105–153. Springer International Publishing, 2014. 
5. B. Biggio, G. Fumera, and F. Roli. Pattern recognition systems under attack: Design issues and 
research challenges. Int’l J. Patt. Recogn. Artif. Intell., 2014, In press. 
6. B. Biggio, I. Corona, D. Maiorca, B. Nelson, N. Srndic, P. Laskov, G. Giacinto, and F. Roli. Evasion 
attacks against machine learning at test time. In H. Blockeel et al., editors, European Conf. on 
Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD), 
Part III, volume 8190 of LNCS, pp. 387–402. Springer, 2013. 
7. B. Biggio, B. Nelson, and P. Laskov. Poisoning attacks against support vector machines. In J. 
Langford and J. Pineau, eds, 29th Int’l Conf. on Machine Learning, pp.1807–1814. Omnipress, 2012. 
8. B. Biggio, I. Pillai, S. R. Bulò, D. Ariu, M. Pelillo, and F. Roli. Is data clustering in adversarial settings 
secure? In Proc. 2013 ACM Workshop on Artificial Intell. and Security, AISec ’13, pages 87–98, New 
York, NY, USA, 2013. ACM. 
9. B. Biggio, S. R. Bulò, I. Pillai, M. Mura, E. Z. Mequanint, M. Pelillo, and F. Roli. Poisoning complete-linkage 
hierarchical clustering. In P. Franti, G. Brown, M. Loog, F. Escolano, and M. Pelillo, editors, 
Joint IAPR Int’l Workshop on Structural, Syntactic, and Statistical Patt. Recogn., volume 8621 of LNCS, 
pages 42–52, Joensuu, Finland, 2014. Springer Berlin Heidelberg. 
http://pralab.diee.unica.it 
40

Contenu connexe

Tendances

On Security and Sparsity of Linear Classifiers for Adversarial Settings
On Security and Sparsity of Linear Classifiers for Adversarial SettingsOn Security and Sparsity of Linear Classifiers for Adversarial Settings
On Security and Sparsity of Linear Classifiers for Adversarial SettingsPluribus One
 
Battista Biggio @ S+SSPR2014, Joensuu, Finland -- Poisoning Complete-Linkage ...
Battista Biggio @ S+SSPR2014, Joensuu, Finland -- Poisoning Complete-Linkage ...Battista Biggio @ S+SSPR2014, Joensuu, Finland -- Poisoning Complete-Linkage ...
Battista Biggio @ S+SSPR2014, Joensuu, Finland -- Poisoning Complete-Linkage ...Pluribus One
 
Battista Biggio @ AISec 2014 - Poisoning Behavioral Malware Clustering
Battista Biggio @ AISec 2014 - Poisoning Behavioral Malware ClusteringBattista Biggio @ AISec 2014 - Poisoning Behavioral Malware Clustering
Battista Biggio @ AISec 2014 - Poisoning Behavioral Malware ClusteringPluribus One
 
Adversarial Learning_Rupam Bhattacharya
Adversarial Learning_Rupam BhattacharyaAdversarial Learning_Rupam Bhattacharya
Adversarial Learning_Rupam BhattacharyaRupam Bhattacharya
 
Causative Adversarial Learning
Causative Adversarial LearningCausative Adversarial Learning
Causative Adversarial LearningDavid Dao
 
WILD PATTERNS - Introduction to Adversarial Machine Learning - ITASEC 2019
WILD PATTERNS - Introduction to Adversarial Machine Learning - ITASEC 2019WILD PATTERNS - Introduction to Adversarial Machine Learning - ITASEC 2019
WILD PATTERNS - Introduction to Adversarial Machine Learning - ITASEC 2019Pluribus One
 
adversarial robustness through local linearization
 adversarial robustness through local linearization adversarial robustness through local linearization
adversarial robustness through local linearizationtaeseon ryu
 
Automated In-memory Malware/Rootkit Detection via Binary Analysis and Machin...
Automated In-memory Malware/Rootkit  Detection via Binary Analysis and Machin...Automated In-memory Malware/Rootkit  Detection via Binary Analysis and Machin...
Automated In-memory Malware/Rootkit Detection via Binary Analysis and Machin...Malachi Jones
 
Wild patterns - Ten years after the rise of Adversarial Machine Learning - Ne...
Wild patterns - Ten years after the rise of Adversarial Machine Learning - Ne...Wild patterns - Ten years after the rise of Adversarial Machine Learning - Ne...
Wild patterns - Ten years after the rise of Adversarial Machine Learning - Ne...Pluribus One
 
On the Malware Detection Problem: Challenges & Novel Approaches
On the Malware Detection Problem: Challenges & Novel ApproachesOn the Malware Detection Problem: Challenges & Novel Approaches
On the Malware Detection Problem: Challenges & Novel ApproachesMarcus Botacin
 
PRACTICAL ADVERSARIAL ATTACKS AGAINST CHALLENGING MODELS ENVIRONMENTS - Moust...
PRACTICAL ADVERSARIAL ATTACKS AGAINST CHALLENGING MODELS ENVIRONMENTS - Moust...PRACTICAL ADVERSARIAL ATTACKS AGAINST CHALLENGING MODELS ENVIRONMENTS - Moust...
PRACTICAL ADVERSARIAL ATTACKS AGAINST CHALLENGING MODELS ENVIRONMENTS - Moust...GeekPwn Keen
 
All You Need to Know to Win a Cybersecurity Adversarial Machine Learning Comp...
All You Need to Know to Win a Cybersecurity Adversarial Machine Learning Comp...All You Need to Know to Win a Cybersecurity Adversarial Machine Learning Comp...
All You Need to Know to Win a Cybersecurity Adversarial Machine Learning Comp...Marcus Botacin
 
Self-learning systems for cyber security
Self-learning systems for cyber securitySelf-learning systems for cyber security
Self-learning systems for cyber securityKim Hammar
 
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning - 2019 Int...
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning - 2019 Int...Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning - 2019 Int...
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning - 2019 Int...Pluribus One
 
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning. ICMLC 201...
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning. ICMLC 201...Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning. ICMLC 201...
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning. ICMLC 201...Pluribus One
 
Mozfest 2018 session slides: Let's fool modern A.I. systems with stickers.
Mozfest 2018 session slides: Let's fool modern A.I. systems with stickers.Mozfest 2018 session slides: Let's fool modern A.I. systems with stickers.
Mozfest 2018 session slides: Let's fool modern A.I. systems with stickers.anant90
 
Who is responsible for adversarial defense
Who is responsible for adversarial defenseWho is responsible for adversarial defense
Who is responsible for adversarial defenseKishor Datta Gupta
 

Tendances (20)

On Security and Sparsity of Linear Classifiers for Adversarial Settings
On Security and Sparsity of Linear Classifiers for Adversarial SettingsOn Security and Sparsity of Linear Classifiers for Adversarial Settings
On Security and Sparsity of Linear Classifiers for Adversarial Settings
 
Battista Biggio @ S+SSPR2014, Joensuu, Finland -- Poisoning Complete-Linkage ...
Battista Biggio @ S+SSPR2014, Joensuu, Finland -- Poisoning Complete-Linkage ...Battista Biggio @ S+SSPR2014, Joensuu, Finland -- Poisoning Complete-Linkage ...
Battista Biggio @ S+SSPR2014, Joensuu, Finland -- Poisoning Complete-Linkage ...
 
Battista Biggio @ AISec 2014 - Poisoning Behavioral Malware Clustering
Battista Biggio @ AISec 2014 - Poisoning Behavioral Malware ClusteringBattista Biggio @ AISec 2014 - Poisoning Behavioral Malware Clustering
Battista Biggio @ AISec 2014 - Poisoning Behavioral Malware Clustering
 
Adversarial Learning_Rupam Bhattacharya
Adversarial Learning_Rupam BhattacharyaAdversarial Learning_Rupam Bhattacharya
Adversarial Learning_Rupam Bhattacharya
 
Causative Adversarial Learning
Causative Adversarial LearningCausative Adversarial Learning
Causative Adversarial Learning
 
WILD PATTERNS - Introduction to Adversarial Machine Learning - ITASEC 2019
WILD PATTERNS - Introduction to Adversarial Machine Learning - ITASEC 2019WILD PATTERNS - Introduction to Adversarial Machine Learning - ITASEC 2019
WILD PATTERNS - Introduction to Adversarial Machine Learning - ITASEC 2019
 
adversarial robustness through local linearization
 adversarial robustness through local linearization adversarial robustness through local linearization
adversarial robustness through local linearization
 
Automated In-memory Malware/Rootkit Detection via Binary Analysis and Machin...
Automated In-memory Malware/Rootkit  Detection via Binary Analysis and Machin...Automated In-memory Malware/Rootkit  Detection via Binary Analysis and Machin...
Automated In-memory Malware/Rootkit Detection via Binary Analysis and Machin...
 
Wild patterns - Ten years after the rise of Adversarial Machine Learning - Ne...
Wild patterns - Ten years after the rise of Adversarial Machine Learning - Ne...Wild patterns - Ten years after the rise of Adversarial Machine Learning - Ne...
Wild patterns - Ten years after the rise of Adversarial Machine Learning - Ne...
 
Adversarial ML - Part 2.pdf
Adversarial ML - Part 2.pdfAdversarial ML - Part 2.pdf
Adversarial ML - Part 2.pdf
 
On the Malware Detection Problem: Challenges & Novel Approaches
On the Malware Detection Problem: Challenges & Novel ApproachesOn the Malware Detection Problem: Challenges & Novel Approaches
On the Malware Detection Problem: Challenges & Novel Approaches
 
PRACTICAL ADVERSARIAL ATTACKS AGAINST CHALLENGING MODELS ENVIRONMENTS - Moust...
PRACTICAL ADVERSARIAL ATTACKS AGAINST CHALLENGING MODELS ENVIRONMENTS - Moust...PRACTICAL ADVERSARIAL ATTACKS AGAINST CHALLENGING MODELS ENVIRONMENTS - Moust...
PRACTICAL ADVERSARIAL ATTACKS AGAINST CHALLENGING MODELS ENVIRONMENTS - Moust...
 
Adversarial ML - Part 1.pdf
Adversarial ML - Part 1.pdfAdversarial ML - Part 1.pdf
Adversarial ML - Part 1.pdf
 
All You Need to Know to Win a Cybersecurity Adversarial Machine Learning Comp...
All You Need to Know to Win a Cybersecurity Adversarial Machine Learning Comp...All You Need to Know to Win a Cybersecurity Adversarial Machine Learning Comp...
All You Need to Know to Win a Cybersecurity Adversarial Machine Learning Comp...
 
Self-learning systems for cyber security
Self-learning systems for cyber securitySelf-learning systems for cyber security
Self-learning systems for cyber security
 
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning - 2019 Int...
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning - 2019 Int...Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning - 2019 Int...
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning - 2019 Int...
 
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning. ICMLC 201...
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning. ICMLC 201...Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning. ICMLC 201...
Wild Patterns: A Half-day Tutorial on Adversarial Machine Learning. ICMLC 201...
 
Robustness in deep learning
Robustness in deep learningRobustness in deep learning
Robustness in deep learning
 
Mozfest 2018 session slides: Let's fool modern A.I. systems with stickers.
Mozfest 2018 session slides: Let's fool modern A.I. systems with stickers.Mozfest 2018 session slides: Let's fool modern A.I. systems with stickers.
Mozfest 2018 session slides: Let's fool modern A.I. systems with stickers.
 
Who is responsible for adversarial defense
Who is responsible for adversarial defenseWho is responsible for adversarial defense
Who is responsible for adversarial defense
 

En vedette

Battista Biggio @ ICML2012: "Poisoning attacks against support vector machines"
Battista Biggio @ ICML2012: "Poisoning attacks against support vector machines"Battista Biggio @ ICML2012: "Poisoning attacks against support vector machines"
Battista Biggio @ ICML2012: "Poisoning attacks against support vector machines"Pluribus One
 
Making neural programming architectures generalize via recursion
Making neural programming architectures generalize via recursionMaking neural programming architectures generalize via recursion
Making neural programming architectures generalize via recursionKaty Lee
 
What Makes Great Infographics
What Makes Great InfographicsWhat Makes Great Infographics
What Makes Great InfographicsSlideShare
 
Masters of SlideShare
Masters of SlideShareMasters of SlideShare
Masters of SlideShareKapost
 
STOP! VIEW THIS! 10-Step Checklist When Uploading to Slideshare
STOP! VIEW THIS! 10-Step Checklist When Uploading to SlideshareSTOP! VIEW THIS! 10-Step Checklist When Uploading to Slideshare
STOP! VIEW THIS! 10-Step Checklist When Uploading to SlideshareEmpowered Presentations
 
10 Ways to Win at SlideShare SEO & Presentation Optimization
10 Ways to Win at SlideShare SEO & Presentation Optimization10 Ways to Win at SlideShare SEO & Presentation Optimization
10 Ways to Win at SlideShare SEO & Presentation OptimizationOneupweb
 
How To Get More From SlideShare - Super-Simple Tips For Content Marketing
How To Get More From SlideShare - Super-Simple Tips For Content MarketingHow To Get More From SlideShare - Super-Simple Tips For Content Marketing
How To Get More From SlideShare - Super-Simple Tips For Content MarketingContent Marketing Institute
 
How to Make Awesome SlideShares: Tips & Tricks
How to Make Awesome SlideShares: Tips & TricksHow to Make Awesome SlideShares: Tips & Tricks
How to Make Awesome SlideShares: Tips & TricksSlideShare
 

En vedette (9)

Battista Biggio @ ICML2012: "Poisoning attacks against support vector machines"
Battista Biggio @ ICML2012: "Poisoning attacks against support vector machines"Battista Biggio @ ICML2012: "Poisoning attacks against support vector machines"
Battista Biggio @ ICML2012: "Poisoning attacks against support vector machines"
 
Making neural programming architectures generalize via recursion
Making neural programming architectures generalize via recursionMaking neural programming architectures generalize via recursion
Making neural programming architectures generalize via recursion
 
What Makes Great Infographics
What Makes Great InfographicsWhat Makes Great Infographics
What Makes Great Infographics
 
Masters of SlideShare
Masters of SlideShareMasters of SlideShare
Masters of SlideShare
 
STOP! VIEW THIS! 10-Step Checklist When Uploading to Slideshare
STOP! VIEW THIS! 10-Step Checklist When Uploading to SlideshareSTOP! VIEW THIS! 10-Step Checklist When Uploading to Slideshare
STOP! VIEW THIS! 10-Step Checklist When Uploading to Slideshare
 
You Suck At PowerPoint!
You Suck At PowerPoint!You Suck At PowerPoint!
You Suck At PowerPoint!
 
10 Ways to Win at SlideShare SEO & Presentation Optimization
10 Ways to Win at SlideShare SEO & Presentation Optimization10 Ways to Win at SlideShare SEO & Presentation Optimization
10 Ways to Win at SlideShare SEO & Presentation Optimization
 
How To Get More From SlideShare - Super-Simple Tips For Content Marketing
How To Get More From SlideShare - Super-Simple Tips For Content MarketingHow To Get More From SlideShare - Super-Simple Tips For Content Marketing
How To Get More From SlideShare - Super-Simple Tips For Content Marketing
 
How to Make Awesome SlideShares: Tips & Tricks
How to Make Awesome SlideShares: Tips & TricksHow to Make Awesome SlideShares: Tips & Tricks
How to Make Awesome SlideShares: Tips & Tricks
 

Similaire à Battista Biggio, Invited Keynote @ AISec 2014 - On Learning and Recognition of Secure Patterns

Subverting Machine Learning Detections for fun and profit
Subverting Machine Learning Detections for fun and profitSubverting Machine Learning Detections for fun and profit
Subverting Machine Learning Detections for fun and profitRam Shankar Siva Kumar
 
Metrics for Security Effort Prioritization
Metrics for Security Effort PrioritizationMetrics for Security Effort Prioritization
Metrics for Security Effort PrioritizationChris Theisen
 
[PH-Neutral 0x7db] Exploit Next Generation®
[PH-Neutral 0x7db] Exploit Next Generation®[PH-Neutral 0x7db] Exploit Next Generation®
[PH-Neutral 0x7db] Exploit Next Generation®Nelson Brito
 
Deep learning QuantUniversity meetup
Deep learning QuantUniversity meetupDeep learning QuantUniversity meetup
Deep learning QuantUniversity meetupQuantUniversity
 
DataMind: An e-learning platform for Data Analysis based on R. RBelgium meetu...
DataMind: An e-learning platform for Data Analysis based on R. RBelgium meetu...DataMind: An e-learning platform for Data Analysis based on R. RBelgium meetu...
DataMind: An e-learning platform for Data Analysis based on R. RBelgium meetu...DataMind-slides
 
SoftwareSecurity.ppt
SoftwareSecurity.pptSoftwareSecurity.ppt
SoftwareSecurity.pptssuserfb92ae
 
Bringing Red vs. Blue to Machine Learning
Bringing Red vs. Blue to Machine LearningBringing Red vs. Blue to Machine Learning
Bringing Red vs. Blue to Machine LearningBobby Filar
 
Barga Data Science lecture 7
Barga Data Science lecture 7Barga Data Science lecture 7
Barga Data Science lecture 7Roger Barga
 
Full stack vulnerability management at scale
Full stack vulnerability management at scaleFull stack vulnerability management at scale
Full stack vulnerability management at scaleEoin Keary
 
Software Security : From school to reality and back!
Software Security : From school to reality and back!Software Security : From school to reality and back!
Software Security : From school to reality and back!Peter Hlavaty
 
Cyber Kill Chain vs. Cyber Criminals
Cyber Kill Chain vs. Cyber CriminalsCyber Kill Chain vs. Cyber Criminals
Cyber Kill Chain vs. Cyber CriminalsDavid Sweigert
 
How to Make a Unicorn: Finding Cybersecurity Talent in the Real World (Boston)
How to Make a Unicorn: Finding Cybersecurity Talent in the Real World (Boston)How to Make a Unicorn: Finding Cybersecurity Talent in the Real World (Boston)
How to Make a Unicorn: Finding Cybersecurity Talent in the Real World (Boston)Franklin Mosley
 
AI and ML Skills for the Testing World Tutorial
AI and ML Skills for the Testing World TutorialAI and ML Skills for the Testing World Tutorial
AI and ML Skills for the Testing World TutorialTariq King
 
Securing your web apps before they hurt the organization
Securing your web apps before they hurt the organizationSecuring your web apps before they hurt the organization
Securing your web apps before they hurt the organizationAntonio Fontes
 
Adversarial machine learning for av software
Adversarial machine learning for av softwareAdversarial machine learning for av software
Adversarial machine learning for av softwarejunseok seo
 
Barga Data Science lecture 4
Barga Data Science lecture 4Barga Data Science lecture 4
Barga Data Science lecture 4Roger Barga
 
Skillful scalefull fullstack security in a state of constant flux
Skillful scalefull fullstack security in a state of constant fluxSkillful scalefull fullstack security in a state of constant flux
Skillful scalefull fullstack security in a state of constant fluxEoin Keary
 
Building Your Application Security Data Hub - OWASP AppSecUSA
Building Your Application Security Data Hub - OWASP AppSecUSABuilding Your Application Security Data Hub - OWASP AppSecUSA
Building Your Application Security Data Hub - OWASP AppSecUSADenim Group
 

Similaire à Battista Biggio, Invited Keynote @ AISec 2014 - On Learning and Recognition of Secure Patterns (20)

Subverting Machine Learning Detections for fun and profit
Subverting Machine Learning Detections for fun and profitSubverting Machine Learning Detections for fun and profit
Subverting Machine Learning Detections for fun and profit
 
Metrics for Security Effort Prioritization
Metrics for Security Effort PrioritizationMetrics for Security Effort Prioritization
Metrics for Security Effort Prioritization
 
[PH-Neutral 0x7db] Exploit Next Generation®
[PH-Neutral 0x7db] Exploit Next Generation®[PH-Neutral 0x7db] Exploit Next Generation®
[PH-Neutral 0x7db] Exploit Next Generation®
 
Deep learning QuantUniversity meetup
Deep learning QuantUniversity meetupDeep learning QuantUniversity meetup
Deep learning QuantUniversity meetup
 
OWASP
OWASPOWASP
OWASP
 
DataMind: An e-learning platform for Data Analysis based on R. RBelgium meetu...
DataMind: An e-learning platform for Data Analysis based on R. RBelgium meetu...DataMind: An e-learning platform for Data Analysis based on R. RBelgium meetu...
DataMind: An e-learning platform for Data Analysis based on R. RBelgium meetu...
 
SoftwareSecurity.ppt
SoftwareSecurity.pptSoftwareSecurity.ppt
SoftwareSecurity.ppt
 
Bringing Red vs. Blue to Machine Learning
Bringing Red vs. Blue to Machine LearningBringing Red vs. Blue to Machine Learning
Bringing Red vs. Blue to Machine Learning
 
CYBRScore Course Catalog
CYBRScore Course CatalogCYBRScore Course Catalog
CYBRScore Course Catalog
 
Barga Data Science lecture 7
Barga Data Science lecture 7Barga Data Science lecture 7
Barga Data Science lecture 7
 
Full stack vulnerability management at scale
Full stack vulnerability management at scaleFull stack vulnerability management at scale
Full stack vulnerability management at scale
 
Software Security : From school to reality and back!
Software Security : From school to reality and back!Software Security : From school to reality and back!
Software Security : From school to reality and back!
 
Cyber Kill Chain vs. Cyber Criminals
Cyber Kill Chain vs. Cyber CriminalsCyber Kill Chain vs. Cyber Criminals
Cyber Kill Chain vs. Cyber Criminals
 
How to Make a Unicorn: Finding Cybersecurity Talent in the Real World (Boston)
How to Make a Unicorn: Finding Cybersecurity Talent in the Real World (Boston)How to Make a Unicorn: Finding Cybersecurity Talent in the Real World (Boston)
How to Make a Unicorn: Finding Cybersecurity Talent in the Real World (Boston)
 
AI and ML Skills for the Testing World Tutorial
AI and ML Skills for the Testing World TutorialAI and ML Skills for the Testing World Tutorial
AI and ML Skills for the Testing World Tutorial
 
Securing your web apps before they hurt the organization
Securing your web apps before they hurt the organizationSecuring your web apps before they hurt the organization
Securing your web apps before they hurt the organization
 
Adversarial machine learning for av software
Adversarial machine learning for av softwareAdversarial machine learning for av software
Adversarial machine learning for av software
 
Barga Data Science lecture 4
Barga Data Science lecture 4Barga Data Science lecture 4
Barga Data Science lecture 4
 
Skillful scalefull fullstack security in a state of constant flux
Skillful scalefull fullstack security in a state of constant fluxSkillful scalefull fullstack security in a state of constant flux
Skillful scalefull fullstack security in a state of constant flux
 
Building Your Application Security Data Hub - OWASP AppSecUSA
Building Your Application Security Data Hub - OWASP AppSecUSABuilding Your Application Security Data Hub - OWASP AppSecUSA
Building Your Application Security Data Hub - OWASP AppSecUSA
 

Plus de Pluribus One

Smart Textiles - Prospettive di mercato - Davide Ariu
Smart Textiles - Prospettive di mercato - Davide Ariu Smart Textiles - Prospettive di mercato - Davide Ariu
Smart Textiles - Prospettive di mercato - Davide Ariu Pluribus One
 
Zahid Akhtar - Ph.D. Defense Slides
Zahid Akhtar - Ph.D. Defense SlidesZahid Akhtar - Ph.D. Defense Slides
Zahid Akhtar - Ph.D. Defense SlidesPluribus One
 
Design of robust classifiers for adversarial environments - Systems, Man, and...
Design of robust classifiers for adversarial environments - Systems, Man, and...Design of robust classifiers for adversarial environments - Systems, Man, and...
Design of robust classifiers for adversarial environments - Systems, Man, and...Pluribus One
 
Robustness of multimodal biometric verification systems under realistic spoof...
Robustness of multimodal biometric verification systems under realistic spoof...Robustness of multimodal biometric verification systems under realistic spoof...
Robustness of multimodal biometric verification systems under realistic spoof...Pluribus One
 
Support Vector Machines Under Adversarial Label Noise (ACML 2011) - Battista ...
Support Vector Machines Under Adversarial Label Noise (ACML 2011) - Battista ...Support Vector Machines Under Adversarial Label Noise (ACML 2011) - Battista ...
Support Vector Machines Under Adversarial Label Noise (ACML 2011) - Battista ...Pluribus One
 
Understanding the risk factors of learning in adversarial environments
Understanding the risk factors of learning in adversarial environmentsUnderstanding the risk factors of learning in adversarial environments
Understanding the risk factors of learning in adversarial environmentsPluribus One
 
Amilab IJCB 2011 Poster
Amilab IJCB 2011 PosterAmilab IJCB 2011 Poster
Amilab IJCB 2011 PosterPluribus One
 
Ariu - Workshop on Artificial Intelligence and Security - 2011
Ariu - Workshop on Artificial Intelligence and Security - 2011Ariu - Workshop on Artificial Intelligence and Security - 2011
Ariu - Workshop on Artificial Intelligence and Security - 2011Pluribus One
 
Ariu - Workshop on Applications of Pattern Analysis 2010 - Poster
Ariu - Workshop on Applications of Pattern Analysis 2010 - PosterAriu - Workshop on Applications of Pattern Analysis 2010 - Poster
Ariu - Workshop on Applications of Pattern Analysis 2010 - PosterPluribus One
 
Ariu - Workshop on Multiple Classifier Systems - 2011
Ariu - Workshop on Multiple Classifier Systems - 2011Ariu - Workshop on Multiple Classifier Systems - 2011
Ariu - Workshop on Multiple Classifier Systems - 2011Pluribus One
 
Ariu - Workshop on Applications of Pattern Analysis
Ariu - Workshop on Applications of Pattern AnalysisAriu - Workshop on Applications of Pattern Analysis
Ariu - Workshop on Applications of Pattern AnalysisPluribus One
 
Ariu - Workshop on Multiple Classifier Systems 2011
Ariu - Workshop on Multiple Classifier Systems 2011Ariu - Workshop on Multiple Classifier Systems 2011
Ariu - Workshop on Multiple Classifier Systems 2011Pluribus One
 
Robustness of Multimodal Biometric Systems under Realistic Spoof Attacks agai...
Robustness of Multimodal Biometric Systems under Realistic Spoof Attacks agai...Robustness of Multimodal Biometric Systems under Realistic Spoof Attacks agai...
Robustness of Multimodal Biometric Systems under Realistic Spoof Attacks agai...Pluribus One
 

Plus de Pluribus One (14)

Smart Textiles - Prospettive di mercato - Davide Ariu
Smart Textiles - Prospettive di mercato - Davide Ariu Smart Textiles - Prospettive di mercato - Davide Ariu
Smart Textiles - Prospettive di mercato - Davide Ariu
 
Zahid Akhtar - Ph.D. Defense Slides
Zahid Akhtar - Ph.D. Defense SlidesZahid Akhtar - Ph.D. Defense Slides
Zahid Akhtar - Ph.D. Defense Slides
 
Design of robust classifiers for adversarial environments - Systems, Man, and...
Design of robust classifiers for adversarial environments - Systems, Man, and...Design of robust classifiers for adversarial environments - Systems, Man, and...
Design of robust classifiers for adversarial environments - Systems, Man, and...
 
Robustness of multimodal biometric verification systems under realistic spoof...
Robustness of multimodal biometric verification systems under realistic spoof...Robustness of multimodal biometric verification systems under realistic spoof...
Robustness of multimodal biometric verification systems under realistic spoof...
 
Support Vector Machines Under Adversarial Label Noise (ACML 2011) - Battista ...
Support Vector Machines Under Adversarial Label Noise (ACML 2011) - Battista ...Support Vector Machines Under Adversarial Label Noise (ACML 2011) - Battista ...
Support Vector Machines Under Adversarial Label Noise (ACML 2011) - Battista ...
 
Understanding the risk factors of learning in adversarial environments
Understanding the risk factors of learning in adversarial environmentsUnderstanding the risk factors of learning in adversarial environments
Understanding the risk factors of learning in adversarial environments
 
Amilab IJCB 2011 Poster
Amilab IJCB 2011 PosterAmilab IJCB 2011 Poster
Amilab IJCB 2011 Poster
 
Ariu - Workshop on Artificial Intelligence and Security - 2011
Ariu - Workshop on Artificial Intelligence and Security - 2011Ariu - Workshop on Artificial Intelligence and Security - 2011
Ariu - Workshop on Artificial Intelligence and Security - 2011
 
Ariu - Workshop on Applications of Pattern Analysis 2010 - Poster
Ariu - Workshop on Applications of Pattern Analysis 2010 - PosterAriu - Workshop on Applications of Pattern Analysis 2010 - Poster
Ariu - Workshop on Applications of Pattern Analysis 2010 - Poster
 
Ariu - Workshop on Multiple Classifier Systems - 2011
Ariu - Workshop on Multiple Classifier Systems - 2011Ariu - Workshop on Multiple Classifier Systems - 2011
Ariu - Workshop on Multiple Classifier Systems - 2011
 
Ariu - Workshop on Applications of Pattern Analysis
Ariu - Workshop on Applications of Pattern AnalysisAriu - Workshop on Applications of Pattern Analysis
Ariu - Workshop on Applications of Pattern Analysis
 
Ariu - Workshop on Multiple Classifier Systems 2011
Ariu - Workshop on Multiple Classifier Systems 2011Ariu - Workshop on Multiple Classifier Systems 2011
Ariu - Workshop on Multiple Classifier Systems 2011
 
Robustness of Multimodal Biometric Systems under Realistic Spoof Attacks agai...
Robustness of Multimodal Biometric Systems under Realistic Spoof Attacks agai...Robustness of Multimodal Biometric Systems under Realistic Spoof Attacks agai...
Robustness of Multimodal Biometric Systems under Realistic Spoof Attacks agai...
 
Wiamis2010 poster
Wiamis2010 posterWiamis2010 poster
Wiamis2010 poster
 

Dernier

The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxheathfieldcps1
 
Separation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesSeparation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesFatimaKhan178732
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...EduSkills OECD
 
PSYCHIATRIC History collection FORMAT.pptx
PSYCHIATRIC   History collection FORMAT.pptxPSYCHIATRIC   History collection FORMAT.pptx
PSYCHIATRIC History collection FORMAT.pptxPoojaSen20
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxSayali Powar
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityGeoBlogs
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxGaneshChakor2
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13Steve Thomason
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsanshu789521
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Krashi Coaching
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAssociation for Project Management
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfchloefrazer622
 
URLs and Routing in the Odoo 17 Website App
URLs and Routing in the Odoo 17 Website AppURLs and Routing in the Odoo 17 Website App
URLs and Routing in the Odoo 17 Website AppCeline George
 
Hybridoma Technology ( Production , Purification , and Application )
Hybridoma Technology  ( Production , Purification , and Application  ) Hybridoma Technology  ( Production , Purification , and Application  )
Hybridoma Technology ( Production , Purification , and Application ) Sakshi Ghasle
 
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Celine George
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxpboyjonauth
 
Science 7 - LAND and SEA BREEZE and its Characteristics
Science 7 - LAND and SEA BREEZE and its CharacteristicsScience 7 - LAND and SEA BREEZE and its Characteristics
Science 7 - LAND and SEA BREEZE and its CharacteristicsKarinaGenton
 
MENTAL STATUS EXAMINATION format.docx
MENTAL     STATUS EXAMINATION format.docxMENTAL     STATUS EXAMINATION format.docx
MENTAL STATUS EXAMINATION format.docxPoojaSen20
 

Dernier (20)

The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
Separation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesSeparation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and Actinides
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 
PSYCHIATRIC History collection FORMAT.pptx
PSYCHIATRIC   History collection FORMAT.pptxPSYCHIATRIC   History collection FORMAT.pptx
PSYCHIATRIC History collection FORMAT.pptx
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activity
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptx
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha elections
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across Sectors
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdf
 
URLs and Routing in the Odoo 17 Website App
URLs and Routing in the Odoo 17 Website AppURLs and Routing in the Odoo 17 Website App
URLs and Routing in the Odoo 17 Website App
 
Hybridoma Technology ( Production , Purification , and Application )
Hybridoma Technology  ( Production , Purification , and Application  ) Hybridoma Technology  ( Production , Purification , and Application  )
Hybridoma Technology ( Production , Purification , and Application )
 
Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1
 
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptx
 
Science 7 - LAND and SEA BREEZE and its Characteristics
Science 7 - LAND and SEA BREEZE and its CharacteristicsScience 7 - LAND and SEA BREEZE and its Characteristics
Science 7 - LAND and SEA BREEZE and its Characteristics
 
MENTAL STATUS EXAMINATION format.docx
MENTAL     STATUS EXAMINATION format.docxMENTAL     STATUS EXAMINATION format.docx
MENTAL STATUS EXAMINATION format.docx
 

Battista Biggio, Invited Keynote @ AISec 2014 - On Learning and Recognition of Secure Patterns

  • 1. Pa#ern Recogni-on and Applica-ons Lab On Learning and Recognition of Secure Patterns University of Cagliari, Italy Department of Electrical and Electronic Engineering BaAsta Biggio Dept. Of Electrical and Electronic Engineering University of Cagliari, Italy Sco#sdale, Arizona, AISec 2014 US, Nov., 7 2014
  • 2. Secure Patterns in Nature • Learning of secure patterns is a well-known problem in nature – Mimicry and camouflage – Arms race between predators and preys http://pralab.diee.unica.it 2
  • 3. Secure Patterns in Computer Security • Similar phenomenon in machine learning and computer security – Obfuscation and polymorphism to hide malicious content Spam emails Malware Start 2007 with a bang! Make WBFS YOUR PORTFOLIO’s first winner of the year ..." http://pralab.diee.unica.it <script type="text/javascript" src=”http:// palwas.servehttp.com//ml.php"></ script>" ... var PGuDO0uq19+PGuDO0uq20; EbphZcei=PVqIW5sV.replace(/jTUZZ/ g,"%"); var eWfleJqh=unescape;" Var NxfaGVHq=“pqXdQ23KZril30”;" q9124=this; var SkuyuppD= q9124["WYd1GoGYc2uG1mYGe2YnltY".r eplace(/[Y12WlG:]/g, "")];SkuyuppD.write(eWfleJqh(Ebph Zcei));" ..." 3
  • 4. Arms Race • Adaptation/evolution is crucial to survive! http://pralab.diee.unica.it Attackers Evasion techniques System designers Design of effective countermeasures text analysis on spam emails visual analysis of attached images Image-based spam … 4
  • 5. Machine Learning in Computer Security http://pralab.diee.unica.it 5
  • 6. Design of Learning-based Systems Training phase http://pralab.diee.unica.it xx x x x x x x x x x x x x x x x x1 x2 ... xd pre-­‐processing and feature extrac-on training data (with labels) classifier learning Linear classifiers: assign a weight to each feature and classify a sample based on the sign of its score f (x) = sign(wT x) +1, malicious −1, legitimate "# $ %$ start" bang portfolio winner" year ... university campus" Start 2007 with a bang! Make WBFS YOUR PORTFOLIO’s first winner of the year ..." start" bang portfolio winner" year ... university campus" 1 1" 1" 1" 1" ..." 0" 0" SPAM x start" bang portfolio winner" year ... university campus" w +2 +1" +1" +1" +1" ..." -3" -4" 6
  • 7. Design of Learning-based Systems Test phase http://pralab.diee.unica.it x1 x2 ... xd pre-­‐processing and feature extrac-on test data x x x x x x x x x x x x x x x x x classifica-on and performance evalua-on e.g., classifica-on accuracy Start 2007 with a bang! Make WBFS YOUR PORTFOLIO’s first winner of the year ..." start" bang portfolio winner" year ... university campus" 1 1" 1" 1" 1" ..." 0" 0" start" bang portfolio winner" year ... university campus" +2 +1" +1" +1" +1" ..." -3" -4" +6 > 0, SPAM" (correctly classified) x w Linear classifiers: assign a weight to each feature and classify a sample based on the sign of its score f (x) = sign(wT x) +1, malicious −1, legitimate "# $ %$ 7
  • 8. Can Machine Learning Be Secure? • Problem: how to evade a linear (trained) classifier? Start 2007 with a bang! Make WBFS YOUR PORTFOLIO’s first winner of the year ..." http://pralab.diee.unica.it start" bang portfolio winner" year ... university campus" 1 1" 1" 1" 1" ..." 0" 0" f (x) = sign(wT x) +6 > 0, SPAM" (correctly classified) x start" bang portfolio winner" year ... university campus" w +2 +1" +1" +1" +1" ..." -3" -4" x’ St4rt 2007 with a b4ng! Make WBFS YOUR PORTFOLIO’s first winner of the year ... campus! start" bang portfolio winner" year ... university campus" 0 0! 1" 1" 1" ..." 0" 1! f (x) = sign(wT x) +3 -4 < 0, HAM" (misclassified email) 8
  • 9. Can Machine Learning Be Secure? • Underlying assumption of machine learning techniques – Training and test data are sampled from the same distribution • In practice – The classifier generalizes well from known examples (+ random noise) – … but can not cope with carefully-crafted attacks! • It should be taught how to do that – Explicitly taking into account adversarial data manipulation – Adversarial machine learning • Problem: how can we assess classifier security in a more systematic manner? http://pralab.diee.unica.it (1) M. Barreno, B. Nelson, R. Sears, A. D. Joseph, and J. D. Tygar. Can machine learning be secure? ASIACCS 2006 (2) B. Biggio, G. Fumera, F. Roli. Security evaluation of pattern classifiers under attack. IEEE Trans. on Knowl. and Data Engineering, 2014 9
  • 10. Security Evalua>on of Pa@ern Classifiers http://pralab.diee.unica.it (1) B. Biggio, G. Fumera, F. Roli. Security evaluation of pattern classifiers under attack. IEEE Trans. on Knowl. and Data Engineering, 2014 (2) B. Biggio et al., Security evaluation of SVMs. SVM applications. Springer, 2014 10
  • 11. Security Evaluation of Pattern Classifiers [B. Biggio, G. Fumera, F. Roli, IEEE Trans. KDE 2014] Adversary model • Goal of the attack • Knowledge of the attacked system • Capability of manipulating data • Attack strategy as an optimization problem Performance of more secure classifiers should degrade more gracefully under attack C2 http://pralab.diee.unica.it Bounded adversary! C1 accuracy 0 performance degradation of text classifiers in spam filtering against a different number of modified words in spam emails Security Evaluation Curve Bound on knowledge / capability (e.g., number of modified words) 11
  • 12. Adversary’s Goal 1. Security violation [Barreno et al. ASIACCS06] – Integrity: evade detection without compromising system operation – Availability: classification errors to compromise system operation – Privacy: gaining confidential information about system users 2. Attack’s specificity [Barreno et al. ASIACCS06] – Targeted/Indiscriminate: misclassification of a specific set/any sample http://pralab.diee.unica.it Integrity Availability Privacy Spearphishing vs phishing 12
  • 13. Adversary’s Knowledge TRAINING DATA • Perfect knowledge - Learning algorithm - Parameters (e.g. feature weights) - Feedback on decisions x x – upper bound on the performance degradation under attack http://pralab.diee.unica.it FEATURE REPRESENTATION LEARNING ALGORITHM e.g., SVM x1 x2 ... xd xx x x x x x x x x x x x x x 13
  • 14. Adversary’s Capability • Attack’s influence [Barreno et al. ASIACCS06] – Manipulation of training/test data • Constraints on data manipulation – maximum number of samples that can be added to the training data • the attacker usually controls only a small fraction of the training samples – maximum amount of modifications • application-specific constraints in feature space • e.g., max. number of words that are modified in spam emails http://pralab.diee.unica.it x ' d(x, x!) ≤ dmax x2 x1 f(x) x Feasible domain 14
  • 15. Main Attack Scenarios • Evasion attacks – Goal: integrity violation, indiscriminate attack – Knowledge: perfect / limited – Capability: manipulating test samples e.g., manipulation of spam emails at test time to evade detection • Poisoning attacks – Goal: availability violation, indiscriminate attack – Knowledge: perfect / limited – Capability: injecting samples into the training data e.g., send spam with some ‘good words’ to poison the anti-spam filter, which may subsequently misclassify legitimate emails containing such ‘good words’ http://pralab.diee.unica.it 15
  • 16. Targeted classifier: SVM • Maximum-margin linear classifier f (x) = sign(g(x)), g(x) = wT x + b min w,b 1 2 http://pralab.diee.unica.it Σ wTw+C max(0,1− yi f (xi )) i 1/margin classifica-on error on training data (hinge loss) 16
  • 17. Kernels and Nonlinearity Σ →g(x) = αi w = αi Σ + b • Enables learning and classification using only dot products between samples • Kernel functions for nonlinear classification – e.g., RBF −2 Kernel ( 2 ) −1.5 −1 −0.5 0 0.5 1 1.5 http://pralab.diee.unica.it yixi i yi x, xi i support vectors k(x, xi ) = exp −γ x − xi 17
  • 18. Evasion A@acks http://pralab.diee.unica.it 18 1. B. Biggio, I. Corona, D. Maiorca, B. Nelson, N. Srndic, P. Laskov, G. Giacinto, and F. Roli. Evasion attacks against machine learning at test time. ECML PKDD, 2013. 2. B. Biggio et al., Security evaluation of SVMs. SVM applications. Springer, 2014
  • 19. A Simple Example • Problem: how to evade a linear (trained) classifier? – We have seen this already… St4rt 2007 with a b4ng! Make WBFS YOUR PORTFOLIO’s first winner of the year ... campus! • But… what if the classifier is nonlinear? f (x) = sign(wT x) – Decision functions can be arbitrarily complicated, with no clear relationship between features (x) and classifier parameters (w) http://pralab.diee.unica.it start" bang portfolio winner" year ... university campus" 0 0! 1" 1" 1" ..." 0" 1! start" bang portfolio winner" year ... university campus" +2 +1" +1" +1" +1" ..." -3" -4" +3 -4 < 0, HAM" (misclassified email) x’ w 19
  • 20. Gradient-descent Evasion Attacks • Goal: maximum-confidence evasion • Knowledge: perfect • Attack strategy: • Non-linear, constrained optimization – Gradient descent: approximate solution for smooth functions • Gradients of g(x) can be analytically computed in many cases – SVMs, Neural networks http://pralab.diee.unica.it −2 −1.5 −1 −0.5 0 0.5 1 x f (x) = sign(g(x)) = +1, malicious −1, legitimate "# $ %$ min x' g(x') s.t. d(x, x') ≤ dmax x ' 20
  • 21. Computing Descent Directions Support vector machines Σ xi )+ b, ∇g(x) = αi g(x) = αi RBF kernel gradient: ∇k(x, xi ) = −2γ exp −γ || x − xi ||2 { }(x − xi ) Neural networks x1 xd http://pralab.diee.unica.it δ1 δk δm xf g(x) w1 wk wm v11 vmd vk1 … … … … yik(x, i yi∇k(x, xi ) Σ i mΣ g(x) = 1+ exp − wkδk (x) k=1 # $ % & ' ( ) * + −1 , - . ∂g(x) ∂xf mΣ = g(x)(1− g(x)) wkδk (x) 1−δk ( (x))vkf k=1 21
  • 22. An Example on Handwritten Digits • Nonlinear SVM (RBF kernel) to discriminate between ‘3’ and ‘7’ • Features: gray-level pixel values – 28 x 28 image = 784 features Before attack (3 vs 7) 5 10 15 20 25 5 10 15 20 25 http://pralab.diee.unica.it After attack, g(x)=0 5 10 15 20 25 5 10 15 20 25 After attack, last iter. 5 10 15 20 25 5 10 15 20 25 0 500 2 1 0 −1 −2 g(x) nNuummbbeerr ooff imteoradtiifoiends gray-level values Few modifications are enough to evade detection! … without even mimicking the targeted class (‘7’) 22
  • 23. Bounding the Adversary’s Knowledge Limited knowledge attacks • Only feature representation and learning algorithm are known • Surrogate data sampled from the same distribution as the classifier’s training data • Classifier’s feedback to label surrogate data PDd(aXta,Y ) http://pralab.diee.unica.it Surrogate training data f(x) Send queries Get labels Learn surrogate classifier f’(x) 23
  • 24. Experiments on PDF Malware Detection • PDF: hierarchy of interconnected objects (keyword/value pairs) • Adversary’s capability – adding up to dmax objects to the PDF – removing objects may compromise the PDF file (and embedded malware code)! http://pralab.diee.unica.it /Type 2 /Page 1 /Encoding 1 … 13 0 obj << /Kids [ 1 0 R 11 0 R ] /Type /Page ... >> end obj 17 0 obj << /Type /Encoding /Differences [ 0 /C0032 ] >> endobj Features: keyword count min x' g(x') s.t. d(x, x') ≤ dmax x ≤ x' 24
  • 25. Experiments on PDF Malware Detection • Dataset: 500 malware samples (Contagio), 500 benign (Internet) – Targeted (surrogate) classifier trained on 500 (100) samples • Evasion rate (FN) at FP=1% vs max. number of added keywords – Averaged on 5 repetitions – Perfect knowledge (PK); Limited knowledge (LK) 1 0.8 0.6 0.4 0.2 0 0 10 20 30 40 50 http://pralab.diee.unica.it d max FN SVM (Linear), λ=0 PK (C=1) LK (C=1) Linear SVM 1 0.8 0.6 0.4 0.2 0 Nonlinear SVM (RBF Kernel) 0 10 20 30 40 50 d max FN SVM (RBF), λ=0 PK (C=1) LK (C=1) Number Number of added keywords to each PDF of added keywords to each PDF 25
  • 26. Poisoning A@acks against SVMs http://pralab.diee.unica.it 26 1. B. Biggio, B. Nelson, P. Laskov. Poisoning attacks against SVMs. ICML, 2012 2. B. Biggio et al., Security evaluation of SVMs. SVM applications. Springer, 2014
  • 27. Poisoning Attacks against SVMs [B. Biggio, B. Nelson, P. Laskov, ICML 2012] malicious h#p://www.vulnerablehotel.com/components/ com_hbssearch/longDesc.php?h_id=1& id=-­‐2%20union%20select%20concat%28username, 0x3a,password%29%20from%20jos_users-­‐-­‐ http://pralab.diee.unica.it Classifier Web Server Tr HTTP requests h#p://www.vulnerablehotel.com/login/ legitimate ✔ ✔ 27
  • 28. Poisoning Attacks against SVMs [B. Biggio, B. Nelson, P. Laskov, ICML 2012] • Poisoning classifier to cause denial of service ✖ malicious legitimate h#p://www.vulnerablehotel.com/login/ h#p://www.vulnerablehotel.com/components/ com_hbssearch/longDesc.php?h_id=1& id=-­‐2%20union%20select%20concat%28username, 0x3a,password%29%20from%20jos_users-­‐-­‐ http://pralab.diee.unica.it Classifier Web Server Tr HTTP requests poisoning a#ack ✖ h#p://www.vulnerablehotel.com/login/components/ h#p://www.vulnerablehotel.com/login/longDesc.php?h_id=1& 28
  • 29. Adversary Model and Attack Strategy • Adversary model – Goal: to maximize classification error (availiability, indiscriminate) – Knowledge: perfect knowledge (trained SVM and TR set are known) – Capability: injecting samples into TR • Attack strategy – optimal attack point xc in TR that maximizes classification error http://pralab.diee.unica.it xc classifica-on error = 0.022 classifica- on error = 0.039 29
  • 30. Adversary Model and Attack Strategy • Adversary model – Goal: to maximize classification error (availiability, indiscriminate) – Knowledge: perfect knowledge (trained SVM and TR set are known) – Capability: injecting samples into TR • Attack strategy – optimal attack point xc in TR that maximizes classification error classifica-on error = 0.022 http://pralab.diee.unica.it classifica-on error as a func-on of xc xc 30
  • 31. Poisoning Attack Algorithm • Max. classification error L(xc) w.r.t. xc through gradient ascent • Gradient is not easy to compute – The training point affects the classification function – Details of the derivation are in the paper http://pralab.diee.unica.it (0) xc xc (0) xc xc 1. B. Biggio, B. Nelson, P. Laskov. Poisoning attacks against SVMs. ICML, 2012 31
  • 32. Experiments on the MNIST digits Single-point attack • Linear SVM; 784 features; TR: 100; VAL: 500; TS: about 2000 – ‘0’ is the malicious (attacking) class – ‘4’ is the legitimate (attacked) one xc http://pralab.diee.unica.it (0) xc 32
  • 33. Experiments on the MNIST digits Multiple-point attack • Linear SVM; 784 features; TR: 100; VAL: 500; TS: about 2000 – ‘0’ is the malicious (attacking) class – ‘4’ is the legitimate (attacked) one http://pralab.diee.unica.it 33
  • 34. http://pralab.diee.unica.it A@acking Clustering 34 1. B. Biggio, I. Pillai, S. R. Bulò, D. Ariu, M. Pelillo, and F. Roli. Is data clustering in adversarial settings secure? AISec, 2013 2. B. Biggio, S. R. Bulò, I. Pillai, M. Mura, E. Z. Mequanint, M. Pelillo, and F. Roli. Poisoning complete-linkage hierarchical clustering. S+SSPR, 2014
  • 35. Attacking Clustering • So far, we have considered supervised learning – Training data consisting of samples and class labels • In many applications, labels are not available or costly to obtain – Unsupervised learning • Training data only include samples – no labels! • Malware clustering – To identify variants of existing malware or new malware families http://pralab.diee.unica.it xx x x x x x x x x x x x x x x x x1 x2 ... xd feature extraction (e.g., URL length, num. of parameters, etc.) data collection (honeypots) clustering of malware families (e.g., similar HTTP requests) for each cluster if … then … else … data analysis / countermeasure design (e.g., signature generation) 35
  • 36. Is Data Clustering Secure? • Attackers can poison input data to subvert malware clustering http://pralab.diee.unica.it x x x x x x x x x x x x x x x x x x1 x2 ... xd feature extraction (e.g., URL length, num. of parameters, etc.) data collection (honeypots) clustering of malware families (e.g., similar HTTP requests) for each cluster if … then … else … data analysis / countermeasure design (e.g., signature generation) Well-­‐cra1ed HTTP requests to subvert clustering h#p://www.vulnerablehotel.com/… h#p://www.vulnerablehotel.com/… h#p://www.vulnerablehotel.com/… h#p://www.vulnerablehotel.com/… … is significantly compromised … becomes useless (too many false alarms, low detection rate) 36
  • 37. Our Work • A framework to identify/design attacks against clustering algorithms – Poisoning: add samples to maximally compromise the clustering output – Obfuscation: hide samples within existing clusters • Some clustering algorithms can be very sensitive to poisoning! – single- and complete-linkage hierarchical clustering can be easily compromised by creating heterogeneous clusters • Details on the attack derivation and implementation are in the papers Clustering on untainted data (80 samples) Clustering after adding 10 attack samples http://pralab.diee.unica.it 37 1. B. Biggio et al. Is data clustering in adversarial settings secure? AISec, 2013 2. B. Biggio et al. Poisoning complete-linkage hierarchical clustering. S+SSPR, 2014
  • 38. Conclusions and Future Work • Learning-based systems can be vulnerable to well-crafted, sophisticated attacks devised by skilled attackers – … that exploit specific vulnerabilities of machine learning algorithms! • Future (and ongoing) work – Privacy Attacks – Secure Learning, Clustering and Feature Selection/Reduction http://pralab.diee.unica.it Secure learning algorithms Attacks against learning 38
  • 39. Joint work with … http://pralab.diee.unica.it ? Any questions Thanks for your a#en-on! 39 … and many others
  • 40. References 1. B. Biggio, G. Fumera, and F. Roli. Security evaluation of pattern classifiers under attack. IEEE Trans. on Knowl. and Data Eng., 26(4):984–996, April 2014. 2. M. Barreno, B. Nelson, R. Sears, A. D. Joseph, and J. D. Tygar. Can machine learning be secure? In Proc. ACM Symp. Information, Computer and Comm. Sec., ASIACCS ’06, pages 16–25, New York, NY, USA, 2006. ACM. 3. M. Barreno, B. Nelson, A. Joseph, and J. Tygar. The security of machine learning. Machine Learning, 81:121–148, 2010. 4. B. Biggio, I. Corona, B. Nelson, B. Rubinstein, D. Maiorca, G. Fumera, G. Giacinto, and F. Roli. Security evaluation of support vector machines in adversarial environments. In Y. Ma and G. Guo, eds, Support Vector Machines Applications, pages 105–153. Springer International Publishing, 2014. 5. B. Biggio, G. Fumera, and F. Roli. Pattern recognition systems under attack: Design issues and research challenges. Int’l J. Patt. Recogn. Artif. Intell., 2014, In press. 6. B. Biggio, I. Corona, D. Maiorca, B. Nelson, N. Srndic, P. Laskov, G. Giacinto, and F. Roli. Evasion attacks against machine learning at test time. In H. Blockeel et al., editors, European Conf. on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD), Part III, volume 8190 of LNCS, pp. 387–402. Springer, 2013. 7. B. Biggio, B. Nelson, and P. Laskov. Poisoning attacks against support vector machines. In J. Langford and J. Pineau, eds, 29th Int’l Conf. on Machine Learning, pp.1807–1814. Omnipress, 2012. 8. B. Biggio, I. Pillai, S. R. Bulò, D. Ariu, M. Pelillo, and F. Roli. Is data clustering in adversarial settings secure? In Proc. 2013 ACM Workshop on Artificial Intell. and Security, AISec ’13, pages 87–98, New York, NY, USA, 2013. ACM. 9. B. Biggio, S. R. Bulò, I. Pillai, M. Mura, E. Z. Mequanint, M. Pelillo, and F. Roli. Poisoning complete-linkage hierarchical clustering. In P. Franti, G. Brown, M. Loog, F. Escolano, and M. Pelillo, editors, Joint IAPR Int’l Workshop on Structural, Syntactic, and Statistical Patt. Recogn., volume 8621 of LNCS, pages 42–52, Joensuu, Finland, 2014. Springer Berlin Heidelberg. http://pralab.diee.unica.it 40