SlideShare a Scribd company logo
1 of 23
Download to read offline
Towards a Theoretical
Framework for LCS

Jan Drugowitsch
Dr Alwyn Barry
Overview
Big Question: Why use LCS?

           Aims and Approach
1.
           Function Approximation
2.
           Dynamic Programming
3.
           Classifier Replacement
4.
           Future Work
5.


May 2006              © A.M.Barry and J. Drugowitsch, 2006
                                         Drugowitsch,
Research Aims

      To establish a formal model for
      Accuracy-based LCS
      To use established models, methods
      and notation
      To transfer method knowledge (analysis
      / improvement)
      To create links to other ML approaches
      and disciplines
May 2006          © A.M.Barry and J. Drugowitsch, 2006
                                     Drugowitsch,
Approach
      Reduce LCS to components:
               Function Approximation
           –
               Dynamic Programming
           –
               Classifier Replacement
           –


      Model the components
      Model their interactions
      Verify models by empirical studies
      Extend models from related domains
May 2006                 © A.M.Barry and J. Drugowitsch, 2006
                                            Drugowitsch,
Function Approximation
      Architecture Outline:
               Set of K classifiers, matching states Sk⊆ S
           –
               Every classifier k approximates the value function
           –
               V over Sk. independently.


      Assumptions
               Function V : S → ℜ is time-invariant
           –
               Time-invariant set of classifiers {S1, …, SK}
           –
               Fixed sampling dist. ∏(S) given by π : S → [0, 1]
           –

                                                                      ~

               Minimising MSE: ∑ π (i )(V (i ) − Vk (i )) 2
           –
                                         i∈S k
May 2006                       © A.M.Barry and J. Drugowitsch, 2006
                                                  Drugowitsch,
Function Approximation
      Linear approx. architecture

               Feature vector φ : S → ℜL
           –
               Approximation parameters vector ωk ∈ ℜL
           –
                               ~
               Approximation V k = φ (i )T ωk
           –


                 Averaging classifiers, φ (i) = (1), ∀i∈S
                 Linear estimators, φ (i) = (1, i)′, ∀i∈S



May 2006                   © A.M.Barry and J. Drugowitsch, 2006
                                              Drugowitsch,
Function Approximation
           Approximation to MSE for k at time t :
                                             I S k (im )(V (im ) − ωk ,tφ (im ))
                                        t
                                 1
                                       ∑
                 ε k ,t                                             ′
                          =
                                                                                2

                            ck , t − 1 m = 0

               Matching: I S : S → {0t,1}
           –                            k


               Match count: ck ,t = ∑ I S (im )
           –                                                      k
                                                        m =0



           Since the sequence of states im is dependant
           on π, the MSE is automatically weighted by
           this distribution.

May 2006                           © A.M.Barry and J. Drugowitsch, 2006
                                                      Drugowitsch,
Function Approximation
      Mixing classifiers
               A classifier k covers Sk ⊆ S
           –
               Need to mix estimates to recover V (i )
                                                 ~
           –
               Requires mixing parameter (will be
           –
               ‘accuracy’ ... see later!)
                 K                                                   K
                                     where ψ k : S → {0,1}, ∑ψ k (i ) = 1
       ~
      V (i ) = ∑ψ k (i )ωkφ (i )
                         ′
                 k =1                                                k =1


               ψk(i) = 0 where i ∉ Sk … classifiers that do
           –
               not match do not contribute

May 2006                      © A.M.Barry and J. Drugowitsch, 2006
                                                 Drugowitsch,
Function Approximation
      Matrix notation
           V = (V (1),L, V ( N ))′                    Dk = I S k D
              Lφ (1)′L                              ~
                                                      Vk = Φωk
                       
           Φ=   L                                          ψ k(1),K,0 
             Lφ ( N )′L                                                  
                                                     Ψk =                
                                                                   O
                  π (1),K,0                                0, K ,ψ ( N ) 
                                                                         
                                                                     k
           D=            O       
                  0, K , π ( N ) 
                                 
                                                      ~K           K
                                                              ~
                                                      V = ∑ ΨkVk = ∑ Ψk Φωk
                   I S k (1),K ,0 
                                                             k =1     k =1
           I Sk =                  
                           O
                  
                   0, K, I S ( N ) 
                                    
                                   
                              k




May 2006                         © A.M.Barry and J. Drugowitsch, 2006
                                                    Drugowitsch,
Function Approximation
      Achievements
               Proof that the MSE of a classifier k at time t is the
           –
               normalised sample variance of the value
               functions over the states matched until time t
               Proof that the expression developed for sample-
           –
               based approximated MSE is a suitable
               approximation for the actual MSE, for finite state
               spaces
               Proof of optimality of approximated MSE from the
           –
               vector-space viewpoint (relates classifier error to
               the Principle of Orthogonality)


May 2006                     © A.M.Barry and J. Drugowitsch, 2006
                                                Drugowitsch,
Function Approximation
      Achievements (cont’d)
               Elaboration of algorithms using the model:
           –
                  Steepest Gradient Descent
                  Least Mean Square
                       Show sensitivity to step size and input range hold in LCS
                   –
                  Normalised LMS
                  Kalman Filter
                       … and its inverse co-variance form
                   –
                       Equivalence to Recursive Least Squares
                   –
                       Derivation of computationally cheap implementation
                   –


               For each, derivation of:
           –
                  Stability criteria
                       For LMS, identification of the excess mean squared
                   –
                       estimation error
                  Time constant bounds
May 2006                       © A.M.Barry and J. Drugowitsch, 2006
                                                  Drugowitsch,
Function Approximation
      Achievements (cont’d)
               Investigation:
           –




May 2006                  © A.M.Barry and J. Drugowitsch, 2006
                                             Drugowitsch,
Function Approximation
      How to set the mixing weights ψk(i)?
               Accuracy-based mixing (YCS-like!)
           –
               Divorce mixing for fn. approx. from fitness
           –

                                          I S k (i )ε k−ν
                         ψ k (i ) =      K

                                       ∑ I S p (i )ε pν
                                                     −

                                        p =1


               Investigating power factor ν ∈ℜ+
           –
               Demonstrate, using the model, that we
           –
               obtain the MLE of the mixed classifiers
               when ν = 1
May 2006                   © A.M.Barry and J. Drugowitsch, 2006
                                              Drugowitsch,
Function Approximation
      Mixing - Empirical Results
               Higher ν, more importance to low error
           –
               Foundν = 1 best compromise
           –




    Classifiers: 0..π/2, π/2.. π, 0.. π                Classifiers: 0..0.5, 0.5.. 1, 0.. 1
May 2006                      © A.M.Barry and J. Drugowitsch, 2006
                                                 Drugowitsch,
Dynamic Programming
      Dynamic Programming:
                         ∞ t                                              
           V (i ) = max E ∑ γ r (it , at , it +1 ) | i0 = i, at = µ (it ) 
             *
                     µ
                          t =0                                            

      Bellman’s Equation:
                                (                                        )
             V * (i ) = max E r (i, a, j ) + γV * ( j ) | j = p (i, a)
                         a∈ A



      Applying the DP operator T to value est.
         (TV )(i ) = max E(r (i, a, j ) + γV ( j ) | j = p (i, a ) )
                     a∈ A

May 2006                        © A.M.Barry and J. Drugowitsch, 2006
                                                   Drugowitsch,
Dynamic Programming
      T is a contraction w.r.t. maximum norm
               Converges to unique fixed point V*= TV*
           –

      With function approximation:
               Approximation after every update:
           –
                                            ()
                            ~          ~
                            Vt +1 = f TVt
               Might not converge for more complex
           –
               approximation architectures
      Is LCS function approximation a non-
      expansion w.r.t. maximum norm?
May 2006                 © A.M.Barry and J. Drugowitsch, 2006
                                            Drugowitsch,
Dynamic Programming
      LCS Value Iteration
               Assume:
           –
                       constant population
                       averaging classifiers, φ (i) = (1), ∀i∈S

                                                                                        ~
                                                                               Vt +1 = TVt
               Classifier is approximating:
           –
               Based on minimising MSE:
           –
                       (                  )
                             ~                      ~                              ~~
                                                                        2                          2
               ∑ Vt +1 (i) − Vk ,t +1 (i) = Vt +1 − Vk ,t +1
                                         2
                                                                                = TVt − Vk ,t +1
                                                                        I Sk                       I Sk
               i∈S k




May 2006                         © A.M.Barry and J. Drugowitsch, 2006
                                                    Drugowitsch,
Dynamic Programming
      LCS Value Iteration (cont’d)
               Minimum is orthog. projection on approx:
           –
                                     ~                                 ~
                                     Vk ,t +1 = Π I Sk Vt +1 = Π I Sk TVt
               Need also to determine new mixing weight
           –
               by evaluating the new approx. error:
                                                                            (   )
                              1                      1
                                               ~                        ~
                                                             2                      2
               ε k ,t +1 =             Vt +1 − Vk ,t +1
                                                =             I − Π Dk TVt
                           Tr( I S k )            Tr( I S k )
                                          I Sk                                      I Sk

                                                                  ~
           – ∴the only time dependency is on Vt
                         K
                   ~                           ~
           – Thus: V =
                         ∑ Ψk ,t +1Π I Sk TVt
                    t +1
                                     k =1


May 2006                             © A.M.Barry and J. Drugowitsch, 2006
                                                        Drugowitsch,
Dynamic Programming
      LCS Asynchronous Value Iteration
               Only update Vk for the matched state
           –
               Minimise: ∑ I S (im )((TVm )(im ) − ωk′ ,tφ (im ) )2
                          t
                                       ~
           –
                                         k
                                m =0



               Thus, approx. costs weighted by state dist
           –

                        (                         )
                 t
                          ~                     ~                      2
               ∑
                                            2
                 π (i ) (TV )(i ) − ωkφ (i ) = TV − Φωk
                                     ′
                                                                       Dk
               i∈S k
                ~                ~
               ∴Vk ,t +1 = Π Dk TVt
                       K
               ~                       ~
               Vt +1 = ∑ Ψk ,t +1Π Dk TVt
                       k =1

May 2006                        © A.M.Barry and J. Drugowitsch, 2006
                                                   Drugowitsch,
Dynamic Programming
      Also derived:
               LCS Q-Learning
           –
                 LMS Implementation
                 Kalman Filter Implementation
                 Demonstrate that the weight update uses the
                 normalised form of local gradient descent
               Model-K
                     based Policy Iteration
           –
                ~                   ~
                Vt +1 = ∑ Ψk Π Dk TµVt
                       k =1
               Step-wise Policy Iteration
           –
               TD(λ) is not possible in LCS due to re-
           –
               evaluation of mixing weights
May 2006                       © A.M.Barry and J. Drugowitsch, 2006
                                                  Drugowitsch,
Dynamic Programming
      Is LCS function approximation a non-expansion w.r.t.
      maximum norm?
               We derive that:
           –
                  Constant Mixing: Yes
                  Arbitrary Mixing: Not necessarily
                  Accuracy-based Mixing: non-expansion (dependant upon a
                  conjecture)


      Is LCS function approximation a non-expansion w.r.t.
      weighted norm?
             We know Tµ is a contraction
           –
             We show ΠDk is a non-expansion
           –
           – We demonstrate:
                  Single Classifier: convergence to a fixed pt
                  Disjoint Classifiers: convergence to a fixed pt
                  Arbitrary Mixing: spectral radius can be ≥ 1 even in some
                  fixed weightings
May 2006                         © A.M.Barry and J. Drugowitsch, 2006
                                                    Drugowitsch,
Classifier Replacement
      How can we define the optimal population
      formally?
               What do we want to reach?
           –
               How does the global fitness of one classifier differ
           –
               from its local fitness (i.e. accuracy)?
               What is the quality of a population?
           –

      Methodology
               Examine methods to define the optimal
           –
               population
               Relate to generalised hill-climbing
           –
               Map to other search criteria
           –

May 2006                    © A.M.Barry and J. Drugowitsch, 2006
                                               Drugowitsch,
Future Work
      Formalising and analysing classifier
      replacement
      Interaction of replacement with function
      approximation
      Interaction of replacement with DP
      Handling stochastic environments
      Error bounds and rates of convergence
      …

Big Answer: LCS is an integrative framework,
… But we need the framework definition!

May 2006            © A.M.Barry and J. Drugowitsch, 2006
                                       Drugowitsch,

More Related Content

Similar to Towards a Theoretical Framework for LCS

Regression Theory
Regression TheoryRegression Theory
Regression TheorySSA KPI
 
Munich07 Foils
Munich07 FoilsMunich07 Foils
Munich07 FoilsAntonini
 
CMOS Analog Design Lect 2
CMOS Analog Design   Lect 2CMOS Analog Design   Lect 2
CMOS Analog Design Lect 2carlosgalup
 
On recent improvements in the conic optimizer in MOSEK
On recent improvements in the conic optimizer in MOSEKOn recent improvements in the conic optimizer in MOSEK
On recent improvements in the conic optimizer in MOSEKedadk
 
Classification Theory
Classification TheoryClassification Theory
Classification TheorySSA KPI
 
Regularized Estimation of Spatial Patterns
Regularized Estimation of Spatial PatternsRegularized Estimation of Spatial Patterns
Regularized Estimation of Spatial PatternsWen-Ting Wang
 

Similar to Towards a Theoretical Framework for LCS (6)

Regression Theory
Regression TheoryRegression Theory
Regression Theory
 
Munich07 Foils
Munich07 FoilsMunich07 Foils
Munich07 Foils
 
CMOS Analog Design Lect 2
CMOS Analog Design   Lect 2CMOS Analog Design   Lect 2
CMOS Analog Design Lect 2
 
On recent improvements in the conic optimizer in MOSEK
On recent improvements in the conic optimizer in MOSEKOn recent improvements in the conic optimizer in MOSEK
On recent improvements in the conic optimizer in MOSEK
 
Classification Theory
Classification TheoryClassification Theory
Classification Theory
 
Regularized Estimation of Spatial Patterns
Regularized Estimation of Spatial PatternsRegularized Estimation of Spatial Patterns
Regularized Estimation of Spatial Patterns
 

More from Xavier Llorà

Meandre 2.0 Alpha Preview
Meandre 2.0 Alpha PreviewMeandre 2.0 Alpha Preview
Meandre 2.0 Alpha PreviewXavier Llorà
 
Soaring the Clouds with Meandre
Soaring the Clouds with MeandreSoaring the Clouds with Meandre
Soaring the Clouds with MeandreXavier Llorà
 
From Galapagos to Twitter: Darwin, Natural Selection, and Web 2.0
From Galapagos to Twitter: Darwin, Natural Selection, and Web 2.0From Galapagos to Twitter: Darwin, Natural Selection, and Web 2.0
From Galapagos to Twitter: Darwin, Natural Selection, and Web 2.0Xavier Llorà
 
Large Scale Data Mining using Genetics-Based Machine Learning
Large Scale Data Mining using   Genetics-Based Machine LearningLarge Scale Data Mining using   Genetics-Based Machine Learning
Large Scale Data Mining using Genetics-Based Machine LearningXavier Llorà
 
Data-Intensive Computing for Competent Genetic Algorithms: A Pilot Study us...
Data-Intensive Computing for  Competent Genetic Algorithms:  A Pilot Study us...Data-Intensive Computing for  Competent Genetic Algorithms:  A Pilot Study us...
Data-Intensive Computing for Competent Genetic Algorithms: A Pilot Study us...Xavier Llorà
 
Scalabiltity in GBML, Accuracy-based Michigan Fuzzy LCS, and new Trends
Scalabiltity in GBML, Accuracy-based Michigan Fuzzy LCS, and new TrendsScalabiltity in GBML, Accuracy-based Michigan Fuzzy LCS, and new Trends
Scalabiltity in GBML, Accuracy-based Michigan Fuzzy LCS, and new TrendsXavier Llorà
 
Pittsburgh Learning Classifier Systems for Protein Structure Prediction: Sca...
Pittsburgh Learning Classifier Systems for Protein  Structure Prediction: Sca...Pittsburgh Learning Classifier Systems for Protein  Structure Prediction: Sca...
Pittsburgh Learning Classifier Systems for Protein Structure Prediction: Sca...Xavier Llorà
 
Learning Classifier Systems for Class Imbalance Problems
Learning Classifier Systems  for Class Imbalance  ProblemsLearning Classifier Systems  for Class Imbalance  Problems
Learning Classifier Systems for Class Imbalance ProblemsXavier Llorà
 
A Retrospective Look at A Retrospective Look at Classifier System ResearchCl...
A Retrospective Look at  A Retrospective Look at  Classifier System ResearchCl...A Retrospective Look at  A Retrospective Look at  Classifier System ResearchCl...
A Retrospective Look at A Retrospective Look at Classifier System ResearchCl...Xavier Llorà
 
XCS: Current capabilities and future challenges
XCS: Current capabilities and future  challengesXCS: Current capabilities and future  challenges
XCS: Current capabilities and future challengesXavier Llorà
 
Negative Selection for Algorithm for Anomaly Detection
Negative Selection for Algorithm for Anomaly DetectionNegative Selection for Algorithm for Anomaly Detection
Negative Selection for Algorithm for Anomaly DetectionXavier Llorà
 
Searle, Intentionality, and the Future of Classifier Systems
Searle, Intentionality, and the  Future of Classifier SystemsSearle, Intentionality, and the  Future of Classifier Systems
Searle, Intentionality, and the Future of Classifier SystemsXavier Llorà
 
Computed Prediction: So far, so good. What now?
Computed Prediction:  So far, so good. What now?Computed Prediction:  So far, so good. What now?
Computed Prediction: So far, so good. What now?Xavier Llorà
 
Linkage Learning for Pittsburgh LCS: Making Problems Tractable
Linkage Learning for Pittsburgh LCS: Making Problems TractableLinkage Learning for Pittsburgh LCS: Making Problems Tractable
Linkage Learning for Pittsburgh LCS: Making Problems TractableXavier Llorà
 
Meandre: Semantic-Driven Data-Intensive Flows in the Clouds
Meandre: Semantic-Driven Data-Intensive Flows in the CloudsMeandre: Semantic-Driven Data-Intensive Flows in the Clouds
Meandre: Semantic-Driven Data-Intensive Flows in the CloudsXavier Llorà
 
ZigZag: The Meandring Language
ZigZag: The Meandring LanguageZigZag: The Meandring Language
ZigZag: The Meandring LanguageXavier Llorà
 
HUMIES 2007 Bronze Winner: Towards Better than Human Capability in Diagnosing...
HUMIES 2007 Bronze Winner: Towards Better than Human Capability in Diagnosing...HUMIES 2007 Bronze Winner: Towards Better than Human Capability in Diagnosing...
HUMIES 2007 Bronze Winner: Towards Better than Human Capability in Diagnosing...Xavier Llorà
 
Do not Match, Inherit: Fitness Surrogates for Genetics-Based Machine Learning...
Do not Match, Inherit: Fitness Surrogates for Genetics-Based Machine Learning...Do not Match, Inherit: Fitness Surrogates for Genetics-Based Machine Learning...
Do not Match, Inherit: Fitness Surrogates for Genetics-Based Machine Learning...Xavier Llorà
 
Towards Better than Human Capability in Diagnosing Prostate Cancer Using Infr...
Towards Better than Human Capability in Diagnosing Prostate Cancer Using Infr...Towards Better than Human Capability in Diagnosing Prostate Cancer Using Infr...
Towards Better than Human Capability in Diagnosing Prostate Cancer Using Infr...Xavier Llorà
 

More from Xavier Llorà (20)

Meandre 2.0 Alpha Preview
Meandre 2.0 Alpha PreviewMeandre 2.0 Alpha Preview
Meandre 2.0 Alpha Preview
 
Soaring the Clouds with Meandre
Soaring the Clouds with MeandreSoaring the Clouds with Meandre
Soaring the Clouds with Meandre
 
From Galapagos to Twitter: Darwin, Natural Selection, and Web 2.0
From Galapagos to Twitter: Darwin, Natural Selection, and Web 2.0From Galapagos to Twitter: Darwin, Natural Selection, and Web 2.0
From Galapagos to Twitter: Darwin, Natural Selection, and Web 2.0
 
Large Scale Data Mining using Genetics-Based Machine Learning
Large Scale Data Mining using   Genetics-Based Machine LearningLarge Scale Data Mining using   Genetics-Based Machine Learning
Large Scale Data Mining using Genetics-Based Machine Learning
 
Data-Intensive Computing for Competent Genetic Algorithms: A Pilot Study us...
Data-Intensive Computing for  Competent Genetic Algorithms:  A Pilot Study us...Data-Intensive Computing for  Competent Genetic Algorithms:  A Pilot Study us...
Data-Intensive Computing for Competent Genetic Algorithms: A Pilot Study us...
 
Scalabiltity in GBML, Accuracy-based Michigan Fuzzy LCS, and new Trends
Scalabiltity in GBML, Accuracy-based Michigan Fuzzy LCS, and new TrendsScalabiltity in GBML, Accuracy-based Michigan Fuzzy LCS, and new Trends
Scalabiltity in GBML, Accuracy-based Michigan Fuzzy LCS, and new Trends
 
Pittsburgh Learning Classifier Systems for Protein Structure Prediction: Sca...
Pittsburgh Learning Classifier Systems for Protein  Structure Prediction: Sca...Pittsburgh Learning Classifier Systems for Protein  Structure Prediction: Sca...
Pittsburgh Learning Classifier Systems for Protein Structure Prediction: Sca...
 
Learning Classifier Systems for Class Imbalance Problems
Learning Classifier Systems  for Class Imbalance  ProblemsLearning Classifier Systems  for Class Imbalance  Problems
Learning Classifier Systems for Class Imbalance Problems
 
A Retrospective Look at A Retrospective Look at Classifier System ResearchCl...
A Retrospective Look at  A Retrospective Look at  Classifier System ResearchCl...A Retrospective Look at  A Retrospective Look at  Classifier System ResearchCl...
A Retrospective Look at A Retrospective Look at Classifier System ResearchCl...
 
XCS: Current capabilities and future challenges
XCS: Current capabilities and future  challengesXCS: Current capabilities and future  challenges
XCS: Current capabilities and future challenges
 
Negative Selection for Algorithm for Anomaly Detection
Negative Selection for Algorithm for Anomaly DetectionNegative Selection for Algorithm for Anomaly Detection
Negative Selection for Algorithm for Anomaly Detection
 
Searle, Intentionality, and the Future of Classifier Systems
Searle, Intentionality, and the  Future of Classifier SystemsSearle, Intentionality, and the  Future of Classifier Systems
Searle, Intentionality, and the Future of Classifier Systems
 
Computed Prediction: So far, so good. What now?
Computed Prediction:  So far, so good. What now?Computed Prediction:  So far, so good. What now?
Computed Prediction: So far, so good. What now?
 
NIGEL 2006 welcome
NIGEL 2006 welcomeNIGEL 2006 welcome
NIGEL 2006 welcome
 
Linkage Learning for Pittsburgh LCS: Making Problems Tractable
Linkage Learning for Pittsburgh LCS: Making Problems TractableLinkage Learning for Pittsburgh LCS: Making Problems Tractable
Linkage Learning for Pittsburgh LCS: Making Problems Tractable
 
Meandre: Semantic-Driven Data-Intensive Flows in the Clouds
Meandre: Semantic-Driven Data-Intensive Flows in the CloudsMeandre: Semantic-Driven Data-Intensive Flows in the Clouds
Meandre: Semantic-Driven Data-Intensive Flows in the Clouds
 
ZigZag: The Meandring Language
ZigZag: The Meandring LanguageZigZag: The Meandring Language
ZigZag: The Meandring Language
 
HUMIES 2007 Bronze Winner: Towards Better than Human Capability in Diagnosing...
HUMIES 2007 Bronze Winner: Towards Better than Human Capability in Diagnosing...HUMIES 2007 Bronze Winner: Towards Better than Human Capability in Diagnosing...
HUMIES 2007 Bronze Winner: Towards Better than Human Capability in Diagnosing...
 
Do not Match, Inherit: Fitness Surrogates for Genetics-Based Machine Learning...
Do not Match, Inherit: Fitness Surrogates for Genetics-Based Machine Learning...Do not Match, Inherit: Fitness Surrogates for Genetics-Based Machine Learning...
Do not Match, Inherit: Fitness Surrogates for Genetics-Based Machine Learning...
 
Towards Better than Human Capability in Diagnosing Prostate Cancer Using Infr...
Towards Better than Human Capability in Diagnosing Prostate Cancer Using Infr...Towards Better than Human Capability in Diagnosing Prostate Cancer Using Infr...
Towards Better than Human Capability in Diagnosing Prostate Cancer Using Infr...
 

Recently uploaded

From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Alan Dix
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfEnterprise Knowledge
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxOnBoard
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGSujit Pal
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slidespraypatel2
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxMalak Abu Hammad
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhisoniya singh
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure servicePooja Nehwal
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking MenDelhi Call girls
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 3652toLead Limited
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitecturePixlogix Infotech
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 

Recently uploaded (20)

From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptx
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAG
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slides
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptx
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC Architecture
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 

Towards a Theoretical Framework for LCS

  • 1. Towards a Theoretical Framework for LCS Jan Drugowitsch Dr Alwyn Barry
  • 2. Overview Big Question: Why use LCS? Aims and Approach 1. Function Approximation 2. Dynamic Programming 3. Classifier Replacement 4. Future Work 5. May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 3. Research Aims To establish a formal model for Accuracy-based LCS To use established models, methods and notation To transfer method knowledge (analysis / improvement) To create links to other ML approaches and disciplines May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 4. Approach Reduce LCS to components: Function Approximation – Dynamic Programming – Classifier Replacement – Model the components Model their interactions Verify models by empirical studies Extend models from related domains May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 5. Function Approximation Architecture Outline: Set of K classifiers, matching states Sk⊆ S – Every classifier k approximates the value function – V over Sk. independently. Assumptions Function V : S → ℜ is time-invariant – Time-invariant set of classifiers {S1, …, SK} – Fixed sampling dist. ∏(S) given by π : S → [0, 1] – ~ Minimising MSE: ∑ π (i )(V (i ) − Vk (i )) 2 – i∈S k May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 6. Function Approximation Linear approx. architecture Feature vector φ : S → ℜL – Approximation parameters vector ωk ∈ ℜL – ~ Approximation V k = φ (i )T ωk – Averaging classifiers, φ (i) = (1), ∀i∈S Linear estimators, φ (i) = (1, i)′, ∀i∈S May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 7. Function Approximation Approximation to MSE for k at time t : I S k (im )(V (im ) − ωk ,tφ (im )) t 1 ∑ ε k ,t ′ = 2 ck , t − 1 m = 0 Matching: I S : S → {0t,1} – k Match count: ck ,t = ∑ I S (im ) – k m =0 Since the sequence of states im is dependant on π, the MSE is automatically weighted by this distribution. May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 8. Function Approximation Mixing classifiers A classifier k covers Sk ⊆ S – Need to mix estimates to recover V (i ) ~ – Requires mixing parameter (will be – ‘accuracy’ ... see later!) K K where ψ k : S → {0,1}, ∑ψ k (i ) = 1 ~ V (i ) = ∑ψ k (i )ωkφ (i ) ′ k =1 k =1 ψk(i) = 0 where i ∉ Sk … classifiers that do – not match do not contribute May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 9. Function Approximation Matrix notation V = (V (1),L, V ( N ))′ Dk = I S k D  Lφ (1)′L  ~ Vk = Φωk   Φ= L   ψ k(1),K,0  Lφ ( N )′L     Ψk =   O  π (1),K,0   0, K ,ψ ( N )      k D= O   0, K , π ( N )    ~K K ~ V = ∑ ΨkVk = ∑ Ψk Φωk  I S k (1),K ,0    k =1 k =1 I Sk =   O   0, K, I S ( N )     k May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 10. Function Approximation Achievements Proof that the MSE of a classifier k at time t is the – normalised sample variance of the value functions over the states matched until time t Proof that the expression developed for sample- – based approximated MSE is a suitable approximation for the actual MSE, for finite state spaces Proof of optimality of approximated MSE from the – vector-space viewpoint (relates classifier error to the Principle of Orthogonality) May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 11. Function Approximation Achievements (cont’d) Elaboration of algorithms using the model: – Steepest Gradient Descent Least Mean Square Show sensitivity to step size and input range hold in LCS – Normalised LMS Kalman Filter … and its inverse co-variance form – Equivalence to Recursive Least Squares – Derivation of computationally cheap implementation – For each, derivation of: – Stability criteria For LMS, identification of the excess mean squared – estimation error Time constant bounds May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 12. Function Approximation Achievements (cont’d) Investigation: – May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 13. Function Approximation How to set the mixing weights ψk(i)? Accuracy-based mixing (YCS-like!) – Divorce mixing for fn. approx. from fitness – I S k (i )ε k−ν ψ k (i ) = K ∑ I S p (i )ε pν − p =1 Investigating power factor ν ∈ℜ+ – Demonstrate, using the model, that we – obtain the MLE of the mixed classifiers when ν = 1 May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 14. Function Approximation Mixing - Empirical Results Higher ν, more importance to low error – Foundν = 1 best compromise – Classifiers: 0..π/2, π/2.. π, 0.. π Classifiers: 0..0.5, 0.5.. 1, 0.. 1 May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 15. Dynamic Programming Dynamic Programming: ∞ t  V (i ) = max E ∑ γ r (it , at , it +1 ) | i0 = i, at = µ (it )  * µ  t =0  Bellman’s Equation: ( ) V * (i ) = max E r (i, a, j ) + γV * ( j ) | j = p (i, a) a∈ A Applying the DP operator T to value est. (TV )(i ) = max E(r (i, a, j ) + γV ( j ) | j = p (i, a ) ) a∈ A May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 16. Dynamic Programming T is a contraction w.r.t. maximum norm Converges to unique fixed point V*= TV* – With function approximation: Approximation after every update: – () ~ ~ Vt +1 = f TVt Might not converge for more complex – approximation architectures Is LCS function approximation a non- expansion w.r.t. maximum norm? May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 17. Dynamic Programming LCS Value Iteration Assume: – constant population averaging classifiers, φ (i) = (1), ∀i∈S ~ Vt +1 = TVt Classifier is approximating: – Based on minimising MSE: – ( ) ~ ~ ~~ 2 2 ∑ Vt +1 (i) − Vk ,t +1 (i) = Vt +1 − Vk ,t +1 2 = TVt − Vk ,t +1 I Sk I Sk i∈S k May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 18. Dynamic Programming LCS Value Iteration (cont’d) Minimum is orthog. projection on approx: – ~ ~ Vk ,t +1 = Π I Sk Vt +1 = Π I Sk TVt Need also to determine new mixing weight – by evaluating the new approx. error: ( ) 1 1 ~ ~ 2 2 ε k ,t +1 = Vt +1 − Vk ,t +1 = I − Π Dk TVt Tr( I S k ) Tr( I S k ) I Sk I Sk ~ – ∴the only time dependency is on Vt K ~ ~ – Thus: V = ∑ Ψk ,t +1Π I Sk TVt t +1 k =1 May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 19. Dynamic Programming LCS Asynchronous Value Iteration Only update Vk for the matched state – Minimise: ∑ I S (im )((TVm )(im ) − ωk′ ,tφ (im ) )2 t ~ – k m =0 Thus, approx. costs weighted by state dist – ( ) t ~ ~ 2 ∑ 2 π (i ) (TV )(i ) − ωkφ (i ) = TV − Φωk ′ Dk i∈S k ~ ~ ∴Vk ,t +1 = Π Dk TVt K ~ ~ Vt +1 = ∑ Ψk ,t +1Π Dk TVt k =1 May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 20. Dynamic Programming Also derived: LCS Q-Learning – LMS Implementation Kalman Filter Implementation Demonstrate that the weight update uses the normalised form of local gradient descent Model-K based Policy Iteration – ~ ~ Vt +1 = ∑ Ψk Π Dk TµVt k =1 Step-wise Policy Iteration – TD(λ) is not possible in LCS due to re- – evaluation of mixing weights May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 21. Dynamic Programming Is LCS function approximation a non-expansion w.r.t. maximum norm? We derive that: – Constant Mixing: Yes Arbitrary Mixing: Not necessarily Accuracy-based Mixing: non-expansion (dependant upon a conjecture) Is LCS function approximation a non-expansion w.r.t. weighted norm? We know Tµ is a contraction – We show ΠDk is a non-expansion – – We demonstrate: Single Classifier: convergence to a fixed pt Disjoint Classifiers: convergence to a fixed pt Arbitrary Mixing: spectral radius can be ≥ 1 even in some fixed weightings May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 22. Classifier Replacement How can we define the optimal population formally? What do we want to reach? – How does the global fitness of one classifier differ – from its local fitness (i.e. accuracy)? What is the quality of a population? – Methodology Examine methods to define the optimal – population Relate to generalised hill-climbing – Map to other search criteria – May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,
  • 23. Future Work Formalising and analysing classifier replacement Interaction of replacement with function approximation Interaction of replacement with DP Handling stochastic environments Error bounds and rates of convergence … Big Answer: LCS is an integrative framework, … But we need the framework definition! May 2006 © A.M.Barry and J. Drugowitsch, 2006 Drugowitsch,