SlideShare une entreprise Scribd logo
1  sur  30
Télécharger pour lire hors ligne
MVPA with SpaceNet: sparse structured priors
https://hal.inria.fr/hal-01147731/document
Elvis DOHMATOB
(Joint work with: M. EICKENBERG, B. THIRION, & G. VAROQUAUX)
L R
y=20
-75
-38
0
38
75 x2
x1
1 Introducing the model
2
1 Brain decoding
We are given:
n = # scans; p = number of voxels in mask
design matrix: X ∈ Rn×p
(brain images)
response vector: y ∈ Rn
(external covariates)
Need to predict y on new data.
Linear model assumption: y ≈ X w
We seek to estimate the weights map, w that
ensures best prediction / classification scores
3
1 The need for regularization
ill-posed problem: high-dimensional (n p)
Typically n ∼ 10 − 103
and p ∼ 104
− 106
We need regularization to reduce dimensions and
encode practioner’s priors on the weights w
4
1 Why spatial priors ?
3D spatial gradient (a linear operator)
: w ∈ Rp
−→ ( x w, y w, zw) ∈ Rp×3
penalize image grad w
⇒ regions
Such priors are reasonable
since brain activity is
spatially correlated
more stable maps and more
predictive than unstructured
priors (e.g SVM)
[Hebiri 2011, Michel 2011,
Baldassare 2012, Grosenick 2013,
Gramfort 2013] 5
1 SpaceNet
SpaceNet is a family of “structure + sparsity”
priors for regularizing the models for brain decoding.
SpaceNet generalizes
TV [Michel 2001],
Smooth-Lasso / GraphNet [Hebiri 2011,
Grosenick 2013], and
TV-L1 [Baldassare 2012, Gramfort 2013].
6
2 Methods
7
2 The SpaceNet regularized model
y = X w + “error”
Optimization problem (regularized model):
minimize 1
2 y − Xw 2
2 + penalty(w)
1
2 y − Xw 2
2 is the loss term, and will be different
for squared-loss, logistic loss, ...
8
2 The SpaceNet regularized model
penalty(w) = αΩρ(w), where
Ωρ(w) := ρ w 1 + (1 − ρ)



1
2 w 2
, for GraphNet
w TV , for TV-L1
...
α (0 < α < +∞) is total amount regularization
ρ (0 < ρ ≤ 1) is a mixing constant called the
1-ratio
ρ = 1 for Lasso
9
2 The SpaceNet regularized model
penalty(w) = αΩρ(w), where
Ωρ(w) := ρ w 1 + (1 − ρ)



1
2 w 2
, for GraphNet
w TV , for TV-L1
...
α (0 < α < +∞) is total amount regularization
ρ (0 < ρ ≤ 1) is a mixing constant called the
1-ratio
ρ = 1 for Lasso
Problem is convex, non-smooth, and
heavily-ill-conditioned. 9
2 Interlude: zoom on ISTA-based algorithms
Settings: min f + g; f smooth, g non-smooth
f and g convex, f L-Lipschitz; both f and g
convex
ISTA: O(L f / ) [Daubechies 2004]
Step 1: Gradient descent on f
Step 2: Proximal operator of g
FISTA: O(L f /
√
) [Beck Teboulle 2009]
= ISTA with a “Nesterov acceleration” trick!
10
2 FISTA: Implementation for TV-L1
1e-3
0.10
1e-3
0.25
1e-3
0.50
1e-3
0.75
1e-3
0.90
1e-2
0.10
1e-2
0.25
1e-2
0.50
1e-2
0.75
1e-2
0.90
1e-1
0.10
1e-1
0.25
1e-1
0.50
1e-1
0.75
1e-1
0.90
102
103
Convergencetimeinseconds
α
1 ratio
[DOHMATOB 2014 (PRNI)]
11
2 FISTA: Implementation for GraphNet
Augment X: ˜X := [X cα,ρ ]T
∈ R(n+3p)×p
⇒ ˜Xz(t)
= Xz(t)
+ cα,ρ (z(t)
)
1. Gradient descent step (datafit term):
w(t+1)
← z(t)
− γ˜XT
(˜Xz(t)
− y)
2. Prox step (penalty term):
w(t+1)
← softαργ(w(t+1)
)
3. Nesterov acceleration:
z(t+1)
← (1 + θ(t)
)w(t+1)
− θ(t)
w(t)
Bottleneck: ∼ 80% of runtime spent doing Xz(t)
!
We badly need speedup!
12
2 Speedup via univariate screening
Whereby we detect and remove irrelevant
voxels before optimization problem is even entered!
13
2 XT
y maps: relevant voxels stick-out
L R
y=20
100% brain vol
14
2 XT
y maps: relevant voxels stick-out
L R
y=20
100% brain vol
L R
y=20
50% brain vol
14
2 XT
y maps: relevant voxels stick-out
L R
y=20
100% brain vol
L R
y=20
50% brain vol
L R
y=20
-75
-38
0
38
75
20% brain vol
14
2 XT
y maps: relevant voxels stick-out
L R
y=20
100% brain vol
L R
y=20
50% brain vol
L R
y=20
-75
-38
0
38
75
20% brain vol
The 20% mask has the 3 bright blobs we would
expect to get
... but contains much less voxels ⇒ less run-time
14
2 Our screening heuristic
tp := pth percentile of the vector |XT
y|.
Discard jth voxel if |XT
j y| < tp
L R
y=20
k = 100% voxels
L R
y=20
k = 50% voxels
L R
y=20
-75
-38
0
38
75
k = 20% voxels
Marginal screening [Lee 2014], but without the
(invertibility) restriction k ≤ min(n, p).
The regularization will do the rest... 15
2 Our screening heuristic
See [DOHMATOB 2015 (PRNI)] for a more
detailed exposition of speedup heuristics developed.
16
2 Automatic model selection via cross-validation
regularization parameters:
0 < αL < ... < α3 < α2 < α1 = αmax
mixing constants:
0 < ρM < ... < ρ3 < ρ2 < ρ1 ≤ 1
Thus L × M grid to search over for best
parameters
(α1, ρ1) (α1, ρ2) (α1, ρ3) ... (α1, ρM)
(α2, ρ1) (α2, ρ2) (α2, ρ3) ... (α2, ρM)
(α3, ρ1) (α3, ρ2) (α3, ρ3) ... (α3, ρM)
... ... ... ... ...
(αL, ρ1) (αL, ρ2) (αL, ρL) ... (αL, ρM)
17
2 Automatic model selection via cross-validation
The final model uses average of the the per-fold
best weights maps (bagging)
This bagging strategy ensures more stable and
robust weights maps
18
3 Some experimental results
19
3 Weights: SpaceNet versus SVM
Faces vs objects classification on [Haxby 2001]
Smooth-Lasso weights TV-L1 weights
SVM weights 20
3 Classification scores: SpaceNet versus SVM
21
3 Concluding remarks
SpaceNet enforces both sparsity and structure,
leading to better prediction / classification scores
and more interpretable brain maps.
The code runs in ∼ 15 minutes for “simple”
datasets, and ∼ 30 minutes for very difficult
datasets.
22
3 Concluding remarks
SpaceNet enforces both sparsity and structure,
leading to better prediction / classification scores
and more interpretable brain maps.
The code runs in ∼ 15 minutes for “simple”
datasets, and ∼ 30 minutes for very difficult
datasets.
In the next release, SpaceNet will feature as part
of Nilearn [Abraham et al. 2014]
http://nilearn.github.io.
22
3 Why XT
y maps give a good relevance measure ?
In an orthogonal design, least-squares solution is
ˆwLS = (XT
X)−1
XT
y = (I)−1
XT
y = XT
y
⇒ (intuition) XT
y bears some info on optimal
solution even for general X
23
3 Why XT
y maps give a good relevance measure ?
In an orthogonal design, least-squares solution is
ˆwLS = (XT
X)−1
XT
y = (I)−1
XT
y = XT
y
⇒ (intuition) XT
y bears some info on optimal
solution even for general X
Marginal screening: Set S = indices of top k
voxels j in terms of |XT
j y| values
In [Lee 2014], k ≤ min(n, p), so that
ˆwLS ∼ (XT
S XS)−1
XT
S y
We don’t require invertibility condition
k ≤ min(n, p). Our spatial regularization will do
the rest!
23
3 Why XT
y maps give a good relevance measure ?
In an orthogonal design, least-squares solution is
ˆwLS = (XT
X)−1
XT
y = (I)−1
XT
y = XT
y
⇒ (intuition) XT
y bears some info on optimal
solution even for general X
Marginal screening: Set S = indices of top k
voxels j in terms of |XT
j y| values
In [Lee 2014], k ≤ min(n, p), so that
ˆwLS ∼ (XT
S XS)−1
XT
S y
We don’t require invertibility condition
k ≤ min(n, p). Our spatial regularization will do
the rest!
Lots of screening rules out there: [El Ghaoui
2010, Liu 2014, Wang 2015, Tibshirani 2010,
Fercoq 2015] 23

Contenu connexe

Tendances

Nonconvex Compressed Sensing with the Sum-of-Squares Method
Nonconvex Compressed Sensing with the Sum-of-Squares MethodNonconvex Compressed Sensing with the Sum-of-Squares Method
Nonconvex Compressed Sensing with the Sum-of-Squares MethodTasuku Soma
 
Optimal interval clustering: Application to Bregman clustering and statistica...
Optimal interval clustering: Application to Bregman clustering and statistica...Optimal interval clustering: Application to Bregman clustering and statistica...
Optimal interval clustering: Application to Bregman clustering and statistica...Frank Nielsen
 
Unbiased Bayes for Big Data
Unbiased Bayes for Big DataUnbiased Bayes for Big Data
Unbiased Bayes for Big DataChristian Robert
 
Markov chain monte_carlo_methods_for_machine_learning
Markov chain monte_carlo_methods_for_machine_learningMarkov chain monte_carlo_methods_for_machine_learning
Markov chain monte_carlo_methods_for_machine_learningAndres Mendez-Vazquez
 
Macrocanonical models for texture synthesis
Macrocanonical models for texture synthesisMacrocanonical models for texture synthesis
Macrocanonical models for texture synthesisValentin De Bortoli
 
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)Universitat Politècnica de Catalunya
 

Tendances (20)

Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Nonconvex Compressed Sensing with the Sum-of-Squares Method
Nonconvex Compressed Sensing with the Sum-of-Squares MethodNonconvex Compressed Sensing with the Sum-of-Squares Method
Nonconvex Compressed Sensing with the Sum-of-Squares Method
 
Jere Koskela slides
Jere Koskela slidesJere Koskela slides
Jere Koskela slides
 
Optimal interval clustering: Application to Bregman clustering and statistica...
Optimal interval clustering: Application to Bregman clustering and statistica...Optimal interval clustering: Application to Bregman clustering and statistica...
Optimal interval clustering: Application to Bregman clustering and statistica...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...
CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...
CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...
 
Distributed ADMM
Distributed ADMMDistributed ADMM
Distributed ADMM
 
QMC: Transition Workshop - Density Estimation by Randomized Quasi-Monte Carlo...
QMC: Transition Workshop - Density Estimation by Randomized Quasi-Monte Carlo...QMC: Transition Workshop - Density Estimation by Randomized Quasi-Monte Carlo...
QMC: Transition Workshop - Density Estimation by Randomized Quasi-Monte Carlo...
 
Unbiased Bayes for Big Data
Unbiased Bayes for Big DataUnbiased Bayes for Big Data
Unbiased Bayes for Big Data
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Markov chain monte_carlo_methods_for_machine_learning
Markov chain monte_carlo_methods_for_machine_learningMarkov chain monte_carlo_methods_for_machine_learning
Markov chain monte_carlo_methods_for_machine_learning
 
Macrocanonical models for texture synthesis
Macrocanonical models for texture synthesisMacrocanonical models for texture synthesis
Macrocanonical models for texture synthesis
 
SASA 2016
SASA 2016SASA 2016
SASA 2016
 
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
 

En vedette

CLOSE PROTECTION
CLOSE PROTECTIONCLOSE PROTECTION
CLOSE PROTECTIONMlu Tolo
 
Batch control system project for a pharmaceutical plant
Batch control system project for a pharmaceutical plantBatch control system project for a pharmaceutical plant
Batch control system project for a pharmaceutical plantISA Interchange
 
BMW ADVANCE 1
BMW ADVANCE 1BMW ADVANCE 1
BMW ADVANCE 1Mlu Tolo
 
Rigid motions & Homogeneous Transformation
Rigid motions & Homogeneous TransformationRigid motions & Homogeneous Transformation
Rigid motions & Homogeneous TransformationShadi Nabil Albarqouni
 
Adobe Premier - R.D.Sivakumar
Adobe Premier - R.D.SivakumarAdobe Premier - R.D.Sivakumar
Adobe Premier - R.D.SivakumarSivakumar R D .
 
Prevención de fenómenos naturales
Prevención de fenómenos naturales Prevención de fenómenos naturales
Prevención de fenómenos naturales Hannita Lizbetita
 
Robust control of speed and temperature in a power plant gas turbine
Robust control of speed and temperature in a power plant gas turbineRobust control of speed and temperature in a power plant gas turbine
Robust control of speed and temperature in a power plant gas turbineISA Interchange
 

En vedette (10)

Algeria
AlgeriaAlgeria
Algeria
 
CLOSE PROTECTION
CLOSE PROTECTIONCLOSE PROTECTION
CLOSE PROTECTION
 
Batch control system project for a pharmaceutical plant
Batch control system project for a pharmaceutical plantBatch control system project for a pharmaceutical plant
Batch control system project for a pharmaceutical plant
 
BMW ADVANCE 1
BMW ADVANCE 1BMW ADVANCE 1
BMW ADVANCE 1
 
Rigid motions & Homogeneous Transformation
Rigid motions & Homogeneous TransformationRigid motions & Homogeneous Transformation
Rigid motions & Homogeneous Transformation
 
Medical robots history
Medical robots historyMedical robots history
Medical robots history
 
DIP Course Projects (HCR)
DIP Course Projects (HCR)DIP Course Projects (HCR)
DIP Course Projects (HCR)
 
Adobe Premier - R.D.Sivakumar
Adobe Premier - R.D.SivakumarAdobe Premier - R.D.Sivakumar
Adobe Premier - R.D.Sivakumar
 
Prevención de fenómenos naturales
Prevención de fenómenos naturales Prevención de fenómenos naturales
Prevención de fenómenos naturales
 
Robust control of speed and temperature in a power plant gas turbine
Robust control of speed and temperature in a power plant gas turbineRobust control of speed and temperature in a power plant gas turbine
Robust control of speed and temperature in a power plant gas turbine
 

Similaire à MVPA with SpaceNet: sparse structured priors

Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3Fabian Pedregosa
 
The world of loss function
The world of loss functionThe world of loss function
The world of loss function홍배 김
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Alexander Litvinenko
 
Digital Signal Processing[ECEG-3171]-Ch1_L03
Digital Signal Processing[ECEG-3171]-Ch1_L03Digital Signal Processing[ECEG-3171]-Ch1_L03
Digital Signal Processing[ECEG-3171]-Ch1_L03Rediet Moges
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Valentin De Bortoli
 
EC8553 Discrete time signal processing
EC8553 Discrete time signal processing EC8553 Discrete time signal processing
EC8553 Discrete time signal processing ssuser2797e4
 
Tutorial APS 2023: Phase transition for statistical estimation: algorithms an...
Tutorial APS 2023: Phase transition for statistical estimation: algorithms an...Tutorial APS 2023: Phase transition for statistical estimation: algorithms an...
Tutorial APS 2023: Phase transition for statistical estimation: algorithms an...Marc Lelarge
 
2 random variables notes 2p3
2 random variables notes 2p32 random variables notes 2p3
2 random variables notes 2p3MuhannadSaleh
 
From RNN to neural networks for cyclic undirected graphs
From RNN to neural networks for cyclic undirected graphsFrom RNN to neural networks for cyclic undirected graphs
From RNN to neural networks for cyclic undirected graphstuxette
 
Bayesian inference on mixtures
Bayesian inference on mixturesBayesian inference on mixtures
Bayesian inference on mixturesChristian Robert
 
Pydata Katya Vasilaky
Pydata Katya VasilakyPydata Katya Vasilaky
Pydata Katya Vasilakyknv4
 
Q-Metrics in Theory and Practice
Q-Metrics in Theory and PracticeQ-Metrics in Theory and Practice
Q-Metrics in Theory and PracticeMagdi Mohamed
 

Similaire à MVPA with SpaceNet: sparse structured priors (20)

Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3
 
KAUST_talk_short.pdf
KAUST_talk_short.pdfKAUST_talk_short.pdf
KAUST_talk_short.pdf
 
The world of loss function
The world of loss functionThe world of loss function
The world of loss function
 
Automatic bayesian cubature
Automatic bayesian cubatureAutomatic bayesian cubature
Automatic bayesian cubature
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...
 
Digital Signal Processing[ECEG-3171]-Ch1_L03
Digital Signal Processing[ECEG-3171]-Ch1_L03Digital Signal Processing[ECEG-3171]-Ch1_L03
Digital Signal Processing[ECEG-3171]-Ch1_L03
 
PCA on graph/network
PCA on graph/networkPCA on graph/network
PCA on graph/network
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
 
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
 
EC8553 Discrete time signal processing
EC8553 Discrete time signal processing EC8553 Discrete time signal processing
EC8553 Discrete time signal processing
 
5.n nmodels i
5.n nmodels i5.n nmodels i
5.n nmodels i
 
Statistical Physics Assignment Help
Statistical Physics Assignment HelpStatistical Physics Assignment Help
Statistical Physics Assignment Help
 
Tutorial APS 2023: Phase transition for statistical estimation: algorithms an...
Tutorial APS 2023: Phase transition for statistical estimation: algorithms an...Tutorial APS 2023: Phase transition for statistical estimation: algorithms an...
Tutorial APS 2023: Phase transition for statistical estimation: algorithms an...
 
2 random variables notes 2p3
2 random variables notes 2p32 random variables notes 2p3
2 random variables notes 2p3
 
From RNN to neural networks for cyclic undirected graphs
From RNN to neural networks for cyclic undirected graphsFrom RNN to neural networks for cyclic undirected graphs
From RNN to neural networks for cyclic undirected graphs
 
smtlecture.6
smtlecture.6smtlecture.6
smtlecture.6
 
Bayesian inference on mixtures
Bayesian inference on mixturesBayesian inference on mixtures
Bayesian inference on mixtures
 
Pydata Katya Vasilaky
Pydata Katya VasilakyPydata Katya Vasilaky
Pydata Katya Vasilaky
 
Input analysis
Input analysisInput analysis
Input analysis
 
Q-Metrics in Theory and Practice
Q-Metrics in Theory and PracticeQ-Metrics in Theory and Practice
Q-Metrics in Theory and Practice
 

Dernier

Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.Silpa
 
Factory Acceptance Test( FAT).pptx .
Factory Acceptance Test( FAT).pptx       .Factory Acceptance Test( FAT).pptx       .
Factory Acceptance Test( FAT).pptx .Poonam Aher Patil
 
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...chandars293
 
Bacterial Identification and Classifications
Bacterial Identification and ClassificationsBacterial Identification and Classifications
Bacterial Identification and ClassificationsAreesha Ahmad
 
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Monika Rani
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bSérgio Sacani
 
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxSCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxRizalinePalanog2
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)Areesha Ahmad
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPirithiRaju
 
COST ESTIMATION FOR A RESEARCH PROJECT.pptx
COST ESTIMATION FOR A RESEARCH PROJECT.pptxCOST ESTIMATION FOR A RESEARCH PROJECT.pptx
COST ESTIMATION FOR A RESEARCH PROJECT.pptxFarihaAbdulRasheed
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.Nitya salvi
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)Areesha Ahmad
 
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Servicenishacall1
 
Module for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learningModule for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learninglevieagacer
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticssakshisoni2385
 
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...Mohammad Khajehpour
 
biology HL practice questions IB BIOLOGY
biology HL practice questions IB BIOLOGYbiology HL practice questions IB BIOLOGY
biology HL practice questions IB BIOLOGY1301aanya
 
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...Silpa
 
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLkantirani197
 

Dernier (20)

Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.
 
Factory Acceptance Test( FAT).pptx .
Factory Acceptance Test( FAT).pptx       .Factory Acceptance Test( FAT).pptx       .
Factory Acceptance Test( FAT).pptx .
 
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
 
Bacterial Identification and Classifications
Bacterial Identification and ClassificationsBacterial Identification and Classifications
Bacterial Identification and Classifications
 
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
 
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxSCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
 
COST ESTIMATION FOR A RESEARCH PROJECT.pptx
COST ESTIMATION FOR A RESEARCH PROJECT.pptxCOST ESTIMATION FOR A RESEARCH PROJECT.pptx
COST ESTIMATION FOR A RESEARCH PROJECT.pptx
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)
 
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
 
Module for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learningModule for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learning
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
 
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...
 
biology HL practice questions IB BIOLOGY
biology HL practice questions IB BIOLOGYbiology HL practice questions IB BIOLOGY
biology HL practice questions IB BIOLOGY
 
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...
 
Clean In Place(CIP).pptx .
Clean In Place(CIP).pptx                 .Clean In Place(CIP).pptx                 .
Clean In Place(CIP).pptx .
 
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
 

MVPA with SpaceNet: sparse structured priors

  • 1. MVPA with SpaceNet: sparse structured priors https://hal.inria.fr/hal-01147731/document Elvis DOHMATOB (Joint work with: M. EICKENBERG, B. THIRION, & G. VAROQUAUX) L R y=20 -75 -38 0 38 75 x2 x1
  • 3. 1 Brain decoding We are given: n = # scans; p = number of voxels in mask design matrix: X ∈ Rn×p (brain images) response vector: y ∈ Rn (external covariates) Need to predict y on new data. Linear model assumption: y ≈ X w We seek to estimate the weights map, w that ensures best prediction / classification scores 3
  • 4. 1 The need for regularization ill-posed problem: high-dimensional (n p) Typically n ∼ 10 − 103 and p ∼ 104 − 106 We need regularization to reduce dimensions and encode practioner’s priors on the weights w 4
  • 5. 1 Why spatial priors ? 3D spatial gradient (a linear operator) : w ∈ Rp −→ ( x w, y w, zw) ∈ Rp×3 penalize image grad w ⇒ regions Such priors are reasonable since brain activity is spatially correlated more stable maps and more predictive than unstructured priors (e.g SVM) [Hebiri 2011, Michel 2011, Baldassare 2012, Grosenick 2013, Gramfort 2013] 5
  • 6. 1 SpaceNet SpaceNet is a family of “structure + sparsity” priors for regularizing the models for brain decoding. SpaceNet generalizes TV [Michel 2001], Smooth-Lasso / GraphNet [Hebiri 2011, Grosenick 2013], and TV-L1 [Baldassare 2012, Gramfort 2013]. 6
  • 8. 2 The SpaceNet regularized model y = X w + “error” Optimization problem (regularized model): minimize 1 2 y − Xw 2 2 + penalty(w) 1 2 y − Xw 2 2 is the loss term, and will be different for squared-loss, logistic loss, ... 8
  • 9. 2 The SpaceNet regularized model penalty(w) = αΩρ(w), where Ωρ(w) := ρ w 1 + (1 − ρ)    1 2 w 2 , for GraphNet w TV , for TV-L1 ... α (0 < α < +∞) is total amount regularization ρ (0 < ρ ≤ 1) is a mixing constant called the 1-ratio ρ = 1 for Lasso 9
  • 10. 2 The SpaceNet regularized model penalty(w) = αΩρ(w), where Ωρ(w) := ρ w 1 + (1 − ρ)    1 2 w 2 , for GraphNet w TV , for TV-L1 ... α (0 < α < +∞) is total amount regularization ρ (0 < ρ ≤ 1) is a mixing constant called the 1-ratio ρ = 1 for Lasso Problem is convex, non-smooth, and heavily-ill-conditioned. 9
  • 11. 2 Interlude: zoom on ISTA-based algorithms Settings: min f + g; f smooth, g non-smooth f and g convex, f L-Lipschitz; both f and g convex ISTA: O(L f / ) [Daubechies 2004] Step 1: Gradient descent on f Step 2: Proximal operator of g FISTA: O(L f / √ ) [Beck Teboulle 2009] = ISTA with a “Nesterov acceleration” trick! 10
  • 12. 2 FISTA: Implementation for TV-L1 1e-3 0.10 1e-3 0.25 1e-3 0.50 1e-3 0.75 1e-3 0.90 1e-2 0.10 1e-2 0.25 1e-2 0.50 1e-2 0.75 1e-2 0.90 1e-1 0.10 1e-1 0.25 1e-1 0.50 1e-1 0.75 1e-1 0.90 102 103 Convergencetimeinseconds α 1 ratio [DOHMATOB 2014 (PRNI)] 11
  • 13. 2 FISTA: Implementation for GraphNet Augment X: ˜X := [X cα,ρ ]T ∈ R(n+3p)×p ⇒ ˜Xz(t) = Xz(t) + cα,ρ (z(t) ) 1. Gradient descent step (datafit term): w(t+1) ← z(t) − γ˜XT (˜Xz(t) − y) 2. Prox step (penalty term): w(t+1) ← softαργ(w(t+1) ) 3. Nesterov acceleration: z(t+1) ← (1 + θ(t) )w(t+1) − θ(t) w(t) Bottleneck: ∼ 80% of runtime spent doing Xz(t) ! We badly need speedup! 12
  • 14. 2 Speedup via univariate screening Whereby we detect and remove irrelevant voxels before optimization problem is even entered! 13
  • 15. 2 XT y maps: relevant voxels stick-out L R y=20 100% brain vol 14
  • 16. 2 XT y maps: relevant voxels stick-out L R y=20 100% brain vol L R y=20 50% brain vol 14
  • 17. 2 XT y maps: relevant voxels stick-out L R y=20 100% brain vol L R y=20 50% brain vol L R y=20 -75 -38 0 38 75 20% brain vol 14
  • 18. 2 XT y maps: relevant voxels stick-out L R y=20 100% brain vol L R y=20 50% brain vol L R y=20 -75 -38 0 38 75 20% brain vol The 20% mask has the 3 bright blobs we would expect to get ... but contains much less voxels ⇒ less run-time 14
  • 19. 2 Our screening heuristic tp := pth percentile of the vector |XT y|. Discard jth voxel if |XT j y| < tp L R y=20 k = 100% voxels L R y=20 k = 50% voxels L R y=20 -75 -38 0 38 75 k = 20% voxels Marginal screening [Lee 2014], but without the (invertibility) restriction k ≤ min(n, p). The regularization will do the rest... 15
  • 20. 2 Our screening heuristic See [DOHMATOB 2015 (PRNI)] for a more detailed exposition of speedup heuristics developed. 16
  • 21. 2 Automatic model selection via cross-validation regularization parameters: 0 < αL < ... < α3 < α2 < α1 = αmax mixing constants: 0 < ρM < ... < ρ3 < ρ2 < ρ1 ≤ 1 Thus L × M grid to search over for best parameters (α1, ρ1) (α1, ρ2) (α1, ρ3) ... (α1, ρM) (α2, ρ1) (α2, ρ2) (α2, ρ3) ... (α2, ρM) (α3, ρ1) (α3, ρ2) (α3, ρ3) ... (α3, ρM) ... ... ... ... ... (αL, ρ1) (αL, ρ2) (αL, ρL) ... (αL, ρM) 17
  • 22. 2 Automatic model selection via cross-validation The final model uses average of the the per-fold best weights maps (bagging) This bagging strategy ensures more stable and robust weights maps 18
  • 23. 3 Some experimental results 19
  • 24. 3 Weights: SpaceNet versus SVM Faces vs objects classification on [Haxby 2001] Smooth-Lasso weights TV-L1 weights SVM weights 20
  • 25. 3 Classification scores: SpaceNet versus SVM 21
  • 26. 3 Concluding remarks SpaceNet enforces both sparsity and structure, leading to better prediction / classification scores and more interpretable brain maps. The code runs in ∼ 15 minutes for “simple” datasets, and ∼ 30 minutes for very difficult datasets. 22
  • 27. 3 Concluding remarks SpaceNet enforces both sparsity and structure, leading to better prediction / classification scores and more interpretable brain maps. The code runs in ∼ 15 minutes for “simple” datasets, and ∼ 30 minutes for very difficult datasets. In the next release, SpaceNet will feature as part of Nilearn [Abraham et al. 2014] http://nilearn.github.io. 22
  • 28. 3 Why XT y maps give a good relevance measure ? In an orthogonal design, least-squares solution is ˆwLS = (XT X)−1 XT y = (I)−1 XT y = XT y ⇒ (intuition) XT y bears some info on optimal solution even for general X 23
  • 29. 3 Why XT y maps give a good relevance measure ? In an orthogonal design, least-squares solution is ˆwLS = (XT X)−1 XT y = (I)−1 XT y = XT y ⇒ (intuition) XT y bears some info on optimal solution even for general X Marginal screening: Set S = indices of top k voxels j in terms of |XT j y| values In [Lee 2014], k ≤ min(n, p), so that ˆwLS ∼ (XT S XS)−1 XT S y We don’t require invertibility condition k ≤ min(n, p). Our spatial regularization will do the rest! 23
  • 30. 3 Why XT y maps give a good relevance measure ? In an orthogonal design, least-squares solution is ˆwLS = (XT X)−1 XT y = (I)−1 XT y = XT y ⇒ (intuition) XT y bears some info on optimal solution even for general X Marginal screening: Set S = indices of top k voxels j in terms of |XT j y| values In [Lee 2014], k ≤ min(n, p), so that ˆwLS ∼ (XT S XS)−1 XT S y We don’t require invertibility condition k ≤ min(n, p). Our spatial regularization will do the rest! Lots of screening rules out there: [El Ghaoui 2010, Liu 2014, Wang 2015, Tibshirani 2010, Fercoq 2015] 23