SlideShare une entreprise Scribd logo
1  sur  37
Télécharger pour lire hors ligne
The promise
                The puzzles
               The prospects




Minimum message length as a truth-conducive
           simplicity measure

                Stephen Petersen
           steve@stevepetersen.net
               Department of Philosophy
                  Niagara University


          Formal Epistemology Workshop
                  2 June 2007



              Steve Petersen   MML and simplicity
The promise
                      The puzzles
                     The prospects




“ How easy it is for the gullible to clutch at the hope that
  somehow, deep in the cabalistic mysteries of information,
  computation, G¨del, Escher, and Bach, there really is an
                   o
  occult connection between simplicity and reality that will
  direct us unswervingly to the Truth; that prior probabilities
  constructed to favor computationally compressible strings
  really are informative and that learning can be defined as
  data-compression. After all, aren’t these things constructed
  out of “information”? I know whereof I speak—I have met
  these glassy-eyed wretches (full professors, even) and they are
  beyond salvation. ”
                                      —Kevin Kelly


                     Steve Petersen      MML and simplicity
The promise
                          The puzzles
                         The prospects


Introduction



      The problem area
          philosophy of science
          epistemology
          cognitive science
      Prospects for the Minimum Message Length (MML) algorithm
      A purported truth-conducive simplicity measure




                         Steve Petersen   MML and simplicity
The promise
                          The puzzles
                         The prospects


Outline
      MML: the promise
  1
       Background
       MML & Bayes
       MML & Kolmogorov
      MML: the puzzles
  2
       MML over Bayes
       UTMs as universal priors
       “Information”
       The data compression analogy
      MML: the prospects
  3
       MML over Bayes
       UTMs as universal priors
       “Information”, data compression, and the nature of abduction

                         Steve Petersen   MML and simplicity
The promise     Background
                         The puzzles     MML & Bayes
                        The prospects    MML & Kolmogorov


Shannon information: the basic idea




      Uncertainty about data source ≈ variety of possible results
      More possibilities ≈ more uncertainty
      Entropy as uncertainty-measure
      Information as reduction in entropy (uncertainty)




                        Steve Petersen   MML and simplicity
The promise        Background
                    The puzzles        MML & Bayes
                   The prospects       MML & Kolmogorov


Shannon information: an example



                             How many yes/no questions needed to
                             find the king?
                             6 divisions of possibility space in half
                                   6
                             (1) =           1
                              2             64
                                        1
                             log 1          = log2 64 = 6
                                       64
                                   2
                             Measure in bits
                             Efficient binary encoding uses 6 digits




                  Steve Petersen       MML and simplicity
The promise     Background
                         The puzzles     MML & Bayes
                        The prospects    MML & Kolmogorov


Shannon information: information and entropy




     Risky approach: e.g., “is it on square f7?”
     If yes, we gain 6 bits of information
     If no, we gain only about .023 bits
     The ideal strategy maximizes information gained with either
     answer




                        Steve Petersen   MML and simplicity
The promise     Background
                         The puzzles     MML & Bayes
                        The prospects    MML & Kolmogorov


Shannon information: subjectivity


                          1
      Suppose odds were       the king is on its original square
                          2
      Clever first question: “is it on its home square?”
      Now on average 3.99 yes/no questions needed
      The entropy now 3.99 bits
      New digital encoding wise, too
      Lower entropy reflects receiver’s lower uncertainty about
      source
      A subjective measure




                        Steve Petersen   MML and simplicity
The promise    Background
                        The puzzles    MML & Bayes
                       The prospects   MML & Kolmogorov


Shannon information: definitions




                                       1
     Shannon information: h(x) = log2 P(x)
     Shannon message length: Lens (x) = h(x)
     Shannon entropy: H(X ) = ∑x P(x)h(x)




                      Steve Petersen   MML and simplicity
The promise     Background
                          The puzzles     MML & Bayes
                         The prospects    MML & Kolmogorov


Kolmogorov complexity in one slide



      Universal Turing Machine U, target string s
      “Description” dU (s): the shortest input to U that returns s
      Kolmogorov complexity KU (s) = |dU (s)|
      Detectable patterns (repeated sequences, the digits of π in
      binary . . . ) → shorter programs
      Interesting measure of randomness
      Lens (s) and K (s) closely related




                         Steve Petersen   MML and simplicity
The promise     Background
                        The puzzles     MML & Bayes
                       The prospects    MML & Kolmogorov


MML: The basic idea




     Bayes: most probable hypothesis given the data
     Shannon: more probable claims allow shorter messages
     MML: find most probable hypothesis by shortest message
     ≈ find truth via simplicity




                       Steve Petersen   MML and simplicity
The promise         Background
                         The puzzles         MML & Bayes
                        The prospects        MML & Kolmogorov


MML is truth-conducive: the “proof”


                              We have truth-related reason to pick

                                                                      P(hi , e)
                                  argmax P(hi |e) = argmax
                                                                       P(e)
                                         i                        i
                                                           = argmax P(hi , e)
                                                                  i
                                                                              1
                                                           = argmin log2
                                                                           P(hi , e)
                                                                  i
                                                           = argmin Lens (hi , e)
                                                                  i

                  1
    Figure: log2 P(x)


                        Steve Petersen       MML and simplicity
The promise     Background
                          The puzzles     MML & Bayes
                         The prospects    MML & Kolmogorov


Simplicity and data-fit

      Epistemic reason to pick shortest statement of hypothesis
      with data
      Big rabbit, small hat
      Note especially:

                Lens (hi , e) = − log2 P(hi , e)
                              = − log2 P(hi )P(e|hi )
                              = − log2 P(hi ) + − log2 P(e|hi )
                              = Lens (hi ) + Lens (e|hi )

      Balance of simplicity and data-fit!


                         Steve Petersen   MML and simplicity
The promise      Background
                    The puzzles      MML & Bayes
                   The prospects     MML & Kolmogorov


The MML explanation



               Table: Wallace’s “explanation”
                Lens (h)             Lens (e|h)
              “assertion”             “details”
                 curve               data error
                pattern                 noise
              regularities           exceptions
              compressor           compressed file




                  Steve Petersen     MML and simplicity
The promise     Background
                           The puzzles     MML & Bayes
                          The prospects    MML & Kolmogorov


Explaining to a UTM



     The eyes get still glassier . . .
     Assume the Shannon-encoding scheme h is computable
     Program a Universal Turing Machine U to compute it with hU
     U determines an implicit prior distribution HU for such
     programs
     Lens (h|HU ) ≈ KU (hU ) = |hU |
     Lens (e|h, HU ) ≈ KU|h (e)




                          Steve Petersen   MML and simplicity
The promise    Background
                        The puzzles    MML & Bayes
                       The prospects   MML & Kolmogorov


UTMs as uninformative Bayesian priors




     Explanation lengths will depend on choice of UTM
     But this difference is bounded above by a constant
     “The” UTM (up to a constant) as a maximally uninformative
     prior
     A step toward a non-subjective information?




                      Steve Petersen   MML and simplicity
The promise     Background
                         The puzzles     MML & Bayes
                        The prospects    MML & Kolmogorov


Algorithmic science




      Wallace suggests we insist U|h also be a UTM
      An “educated” UTM
          Its implicit prior favors h
          Can be further educated with any (computable) hypothesis
      “Normal” and “revolutionary” algorithmic science




                        Steve Petersen   MML and simplicity
The promise     Background
                       The puzzles     MML & Bayes
                      The prospects    MML & Kolmogorov


The robot scientist




                      Steve Petersen   MML and simplicity
MML over Bayes
                          The promise
                                         UTMs as priors
                          The puzzles
                                         “Information”
                         The prospects
                                         The data compression analogy


How is MML any better than Bayesian inference?




  You might reasonably ask:
    “ Why do we need MML? Why don’t we just pick the
      hypothesis with the highest Bayesian posterior probability? ”




                        Steve Petersen   MML and simplicity
MML over Bayes
                          The promise
                                         UTMs as priors
                          The puzzles
                                         “Information”
                         The prospects
                                         The data compression analogy


UTMs as universal priors




      The “up to constant” part is rarely negligible
      Suppose UTM2 emulates UTM1 with just a 272-bit program
      P(h|UTM2 ) = 2−272 · P(h|UTM1 )




                        Steve Petersen   MML and simplicity
MML over Bayes
                          The promise
                                         UTMs as priors
                          The puzzles
                                         “Information”
                         The prospects
                                         The data compression analogy


The notion of “information”



      Kevin Kelly is (rightly) concerned about abuses of
      “information”
      Ambiguity: Shannon-information vs. learning theory
      information
      Compare the ambiguity in a word’s “representing” a thing,
      and a thinker’s “representing” a thing
      The former is “derived”, the latter “original”
      I suspect this is the same ambiguity




                        Steve Petersen   MML and simplicity
MML over Bayes
                          The promise
                                         UTMs as priors
                          The puzzles
                                         “Information”
                         The prospects
                                         The data compression analogy


The null hypothesis


      One might get the impression that Lens (h) + Lens (e|h) should
      be shorter than Lens (e)
      Wallace encourages this with his translation to data
      compression
      Also with his discussion of the null hypothesis:
        “ We will require the length of the explanation message
          which states and uses an acceptable theory to be shorter
          than any message restating the data using only the
          implications of prior premises. ”
      But this is never the case!



                        Steve Petersen   MML and simplicity
MML over Bayes
                         The promise
                                         UTMs as priors
                         The puzzles
                                         “Information”
                        The prospects
                                         The data compression analogy


Data compression is impossible?


      Suppose for contradiction

                       Lens (h) + Lens (e|h) < Lens (e)
                  − log P(h) + − log P(e|h) < − log P(e)
                           − log P(h)P(e|h) < − log P(e)
                                  − log P(h, e) < − log P(e)
                                         P(h, e) > P(e)

      Straightforwardly contradicts probability axioms
      There’s some important disanalogy with data compression



                        Steve Petersen   MML and simplicity
MML over Bayes
                        The promise
                                        UTMs as priors
                        The puzzles
                                        “Information”
                       The prospects
                                        The data compression analogy


Too much information




     Elsewhere Wallace acknowledges this
     The explanation will exceed encoded data length by Lens (h|e)
     Still elsewhere, Wallace takes this as a virtue of MML
     Lens (e, h) > Lens (e) because MML is abductive
     Why send this extra information, given the cost?




                       Steve Petersen   MML and simplicity
The promise    MML over Bayes
                         The puzzles    UTMs as priors
                        The prospects   The nature of abduction


The advantage of MML over Bayes



     For discrete H , MML = Bayesian inference
     MML has the advantage for continuous-valued H
     Bayes: a posterior density, sensitive to parametrization of
     priors
     MML finds the right discretization H ∗
     And the right h ∈ H ∗




                       Steve Petersen   MML and simplicity
The promise    MML over Bayes
                              The puzzles    UTMs as priors
                             The prospects   The nature of abduction


A classic example

  Infer coin bias from a series of flips.

      Table: Potential explanation lenghts (in nits) for 20 heads in 100

         h∈H ∗       |H ∗|      Lens (h)     Lens (e|h)           Total length
                       0           0          51.900                51.900
            0
            /
           .20        100        4.615        50.040                54.655
          .205         99        3.922        50.045                53.970
          .50          1           0          69.315                69.315
           .25         10        1.907        50.161                52.068




                           Steve Petersen    MML and simplicity
The promise    MML over Bayes
                        The puzzles    UTMs as priors
                       The prospects   The nature of abduction


Bayes, MML, and Kolmogorov




     MML’s real strength over Bayesian inference
     MML also gives approximations for Kolmogorov complexity
     A handy bridge between the two
     Neither MML nor Kolmogorov complexity are computable




                      Steve Petersen   MML and simplicity
The promise     MML over Bayes
                         The puzzles     UTMs as priors
                        The prospects    The nature of abduction


The simplest UTM


   “ . . . the complexity of a TM is monotone increasing with its
     number of states. An overly simple TM, say one with only one
     or two states, cannot be universal. There must be, and is, a
     simplest UTM, or a small set of simplest UTMs. Adoption of
     such a UTM as receiver can reasonably be regarded as
     expressing no expectation about the theory or estimate to be
     inferred, save that it will be computable . . . Adoption of any
     UTM (or TM) with more states seems necessary to assume
     something extra, i.e., to adopt a “less ignorant” prior. We
     therefore suggest that the only prior expressing total ignorance
     is that implied by a simplest UTM. ”



                        Steve Petersen   MML and simplicity
The promise    MML over Bayes
                        The puzzles    UTMs as priors
                       The prospects   The nature of abduction


“The” simplest UTM




     Will still depend on the way of specifying UTMs
     Wallace says the “simplest” in any such should do
     Is that right?
     Would an educated “simplest” UTM use “green” and not
     “grue”?




                      Steve Petersen   MML and simplicity
The promise     MML over Bayes
                        The puzzles     UTMs as priors
                       The prospects    The nature of abduction


The simplest UTM?



     Wallace may be wrong that a UTM “with only one or two
     states cannot be universal”
     Wolfram thinks this “two state, three color” one might be:




     $25,000 prize! (http://wolframprize.org)




                       Steve Petersen   MML and simplicity
The promise     MML over Bayes
                          The puzzles     UTMs as priors
                         The prospects    The nature of abduction


Naturalized intentionality



      “Original information” as “original intentionality”
      Naturalized theories of intentionality
  Millikan-Dretske
  When it’s a function of a creature to have an internal element
  covary with aspects of external circumstances, that element can be
  representational if such covariance is supposed to help satisfy a
  need of the creature.




                         Steve Petersen   MML and simplicity
The promise     MML over Bayes
                         The puzzles     UTMs as priors
                        The prospects    The nature of abduction


Robot scientist representing



      Millikan: the “consumer-side” makes some symbol
      representational
      Our robot scientist could gain “original information” via MML
      (Assuming that inferring hypotheses meant to covary with a
      data source can help serve its needs)
      The priors in these needs?




                        Steve Petersen   MML and simplicity
The promise     MML over Bayes
                        The puzzles     UTMs as priors
                       The prospects    The nature of abduction


Solomonoff’s specter


     Again, why abduction?
     Could carry forward the entire posterior distribution over H
     Solomonoff’s “algorithmic probability” predictive strategy
     Wallace: Solomonoff will predict better, but
      “ MML attempts to mimic the discovery of natural laws,
         whereas Solomonoff wants to predict what will happen
         next with no explicit concern for understanding why. ”
     Solomonoff is deductive, MML ampliative




                       Steve Petersen   MML and simplicity
The promise     MML over Bayes
                         The puzzles     UTMs as priors
                        The prospects    The nature of abduction


Why abduction?




     Again, why abduction?
     Especially given cost in message length and predictive power?
     Marcus Hutter’s “universal AI” based on Solomonoff
     Hutter: artificial intelligence ≈ data compression
     50,000 prize!   (http://prize.hutter1.net)




                        Steve Petersen   MML and simplicity
The promise     MML over Bayes
                        The puzzles     UTMs as priors
                       The prospects    The nature of abduction


Abduction and AI


     Hunch: AI as the problem of lossy compression
     Glean the important parts, toss the rest
     “Important” parts agent-relative
     Compare consumer-side intentionality
     Deductive methods remain badly intractable—maybe not
     coincidentally
     MML might help with these (related?) problems
     Abduction as “mere heuristics” of intelligent creatures




                       Steve Petersen   MML and simplicity
The promise     MML over Bayes
                          The puzzles     UTMs as priors
                         The prospects    The nature of abduction


Summary

    MML is not
          A source of non-subjective information
          A clear, clean solution to the problem of the priors
          A miraculous oracle of truth from simplicity
    MML is
          A decent algorithm for Bayesian inference over continuous
          hypothesis spaces
          A possible step in gaining “real” information
          A nice bridge of Bayes and Kolmogorov
          Part of an intriguing model for scientific progress
          Some truth-related vindication of data compression (a form of
          simplicity)



                         Steve Petersen   MML and simplicity
The promise     MML over Bayes
                            The puzzles     UTMs as priors
                           The prospects    The nature of abduction


Thanks


  Thanks to:
      Dennis Whitcomb, Rutgers
      Shahed Sharif, Duke
      Ken Regan, SUNY Buffalo
      Alex Bertland, Niagara University
      Niagara University
      Branden Fitelson and FEW
      You




                           Steve Petersen   MML and simplicity

Contenu connexe

Dernier

Uk-NO1 Amil In Karachi Best Amil In Karachi Bangali Baba In Karachi Aamil In ...
Uk-NO1 Amil In Karachi Best Amil In Karachi Bangali Baba In Karachi Aamil In ...Uk-NO1 Amil In Karachi Best Amil In Karachi Bangali Baba In Karachi Aamil In ...
Uk-NO1 Amil In Karachi Best Amil In Karachi Bangali Baba In Karachi Aamil In ...Amil baba
 
Bald Philosopher, a story for entertainment.docx
Bald Philosopher, a story for entertainment.docxBald Philosopher, a story for entertainment.docx
Bald Philosopher, a story for entertainment.docxazuremorn
 
Flying Avocado Cat Cryptocurrency Created, Coded, Generated and Named by Grok...
Flying Avocado Cat Cryptocurrency Created, Coded, Generated and Named by Grok...Flying Avocado Cat Cryptocurrency Created, Coded, Generated and Named by Grok...
Flying Avocado Cat Cryptocurrency Created, Coded, Generated and Named by Grok...TeslaStakeHolder
 
Princess Jahan's Tuition Classes, a story for entertainment
Princess Jahan's Tuition Classes, a story for entertainmentPrincess Jahan's Tuition Classes, a story for entertainment
Princess Jahan's Tuition Classes, a story for entertainmentazuremorn
 
What Life Would Be Like From A Different Perspective (saltyvixenstories.com)
What Life Would Be Like From A Different Perspective (saltyvixenstories.com)What Life Would Be Like From A Different Perspective (saltyvixenstories.com)
What Life Would Be Like From A Different Perspective (saltyvixenstories.com)Salty Vixen Stories & More
 
Fight Scene Storyboard (Action/Adventure Animation)
Fight Scene Storyboard (Action/Adventure Animation)Fight Scene Storyboard (Action/Adventure Animation)
Fight Scene Storyboard (Action/Adventure Animation)finlaygoodall2
 
THE MEDIC, A STORY for entertainment.docx
THE MEDIC, A STORY for entertainment.docxTHE MEDIC, A STORY for entertainment.docx
THE MEDIC, A STORY for entertainment.docxazuremorn
 
A Spotlight on Darla Leigh Pittman Rodgers: Aaron Rodgers' Mother
A Spotlight on Darla Leigh Pittman Rodgers: Aaron Rodgers' MotherA Spotlight on Darla Leigh Pittman Rodgers: Aaron Rodgers' Mother
A Spotlight on Darla Leigh Pittman Rodgers: Aaron Rodgers' Motherget joys
 
Behind the Scenes The Life of Enzo Zelocchi, a Hollywood Film Producer.pdf
Behind the Scenes The Life of Enzo Zelocchi, a Hollywood Film Producer.pdfBehind the Scenes The Life of Enzo Zelocchi, a Hollywood Film Producer.pdf
Behind the Scenes The Life of Enzo Zelocchi, a Hollywood Film Producer.pdfEnzo Zelocchi Fan Page
 
NO1 Certified kala ilam Expert In Peshwar Kala Jadu Specialist In Peshwar Kal...
NO1 Certified kala ilam Expert In Peshwar Kala Jadu Specialist In Peshwar Kal...NO1 Certified kala ilam Expert In Peshwar Kala Jadu Specialist In Peshwar Kal...
NO1 Certified kala ilam Expert In Peshwar Kala Jadu Specialist In Peshwar Kal...Amil Baba Dawood bangali
 
Biswanath Byam Samiti Open Quiz 2022 by Qui9 Grand Finale
Biswanath Byam Samiti Open Quiz 2022 by Qui9 Grand FinaleBiswanath Byam Samiti Open Quiz 2022 by Qui9 Grand Finale
Biswanath Byam Samiti Open Quiz 2022 by Qui9 Grand FinaleQui9 (Ultimate Quizzing)
 
ECOLUXE pre-ESPYS Ultimate Sports Lounge 2024
ECOLUXE pre-ESPYS Ultimate Sports Lounge 2024ECOLUXE pre-ESPYS Ultimate Sports Lounge 2024
ECOLUXE pre-ESPYS Ultimate Sports Lounge 2024Durkin Entertainment LLC
 
Aesthetic Design Inspiration by Slidesgo.pptx
Aesthetic Design Inspiration by Slidesgo.pptxAesthetic Design Inspiration by Slidesgo.pptx
Aesthetic Design Inspiration by Slidesgo.pptxsayemalkadripial4
 
Statement Of Intent - - Copy.documentfile
Statement Of Intent - - Copy.documentfileStatement Of Intent - - Copy.documentfile
Statement Of Intent - - Copy.documentfilef4ssvxpz62
 
NO1 Certified Black magic specialist,Expert in Pakistan Amil Baba kala ilam E...
NO1 Certified Black magic specialist,Expert in Pakistan Amil Baba kala ilam E...NO1 Certified Black magic specialist,Expert in Pakistan Amil Baba kala ilam E...
NO1 Certified Black magic specialist,Expert in Pakistan Amil Baba kala ilam E...Amil Baba Dawood bangali
 
Taken Pilot Episode Story pitch Document
Taken Pilot Episode Story pitch DocumentTaken Pilot Episode Story pitch Document
Taken Pilot Episode Story pitch Documentf4ssvxpz62
 

Dernier (20)

Uk-NO1 Amil In Karachi Best Amil In Karachi Bangali Baba In Karachi Aamil In ...
Uk-NO1 Amil In Karachi Best Amil In Karachi Bangali Baba In Karachi Aamil In ...Uk-NO1 Amil In Karachi Best Amil In Karachi Bangali Baba In Karachi Aamil In ...
Uk-NO1 Amil In Karachi Best Amil In Karachi Bangali Baba In Karachi Aamil In ...
 
Sincerely, The Friday Club - Farewell Quiz-Finals.pptx
Sincerely, The Friday Club - Farewell Quiz-Finals.pptxSincerely, The Friday Club - Farewell Quiz-Finals.pptx
Sincerely, The Friday Club - Farewell Quiz-Finals.pptx
 
Bald Philosopher, a story for entertainment.docx
Bald Philosopher, a story for entertainment.docxBald Philosopher, a story for entertainment.docx
Bald Philosopher, a story for entertainment.docx
 
Flying Avocado Cat Cryptocurrency Created, Coded, Generated and Named by Grok...
Flying Avocado Cat Cryptocurrency Created, Coded, Generated and Named by Grok...Flying Avocado Cat Cryptocurrency Created, Coded, Generated and Named by Grok...
Flying Avocado Cat Cryptocurrency Created, Coded, Generated and Named by Grok...
 
Princess Jahan's Tuition Classes, a story for entertainment
Princess Jahan's Tuition Classes, a story for entertainmentPrincess Jahan's Tuition Classes, a story for entertainment
Princess Jahan's Tuition Classes, a story for entertainment
 
What Life Would Be Like From A Different Perspective (saltyvixenstories.com)
What Life Would Be Like From A Different Perspective (saltyvixenstories.com)What Life Would Be Like From A Different Perspective (saltyvixenstories.com)
What Life Would Be Like From A Different Perspective (saltyvixenstories.com)
 
Fight Scene Storyboard (Action/Adventure Animation)
Fight Scene Storyboard (Action/Adventure Animation)Fight Scene Storyboard (Action/Adventure Animation)
Fight Scene Storyboard (Action/Adventure Animation)
 
THE MEDIC, A STORY for entertainment.docx
THE MEDIC, A STORY for entertainment.docxTHE MEDIC, A STORY for entertainment.docx
THE MEDIC, A STORY for entertainment.docx
 
S10_E06-Sincerely,The Friday Club- Prelims Farewell Quiz.pptx
S10_E06-Sincerely,The Friday Club- Prelims Farewell Quiz.pptxS10_E06-Sincerely,The Friday Club- Prelims Farewell Quiz.pptx
S10_E06-Sincerely,The Friday Club- Prelims Farewell Quiz.pptx
 
S10_E02_How to Pimp Social Media 101.pptx
S10_E02_How to Pimp Social Media 101.pptxS10_E02_How to Pimp Social Media 101.pptx
S10_E02_How to Pimp Social Media 101.pptx
 
A Spotlight on Darla Leigh Pittman Rodgers: Aaron Rodgers' Mother
A Spotlight on Darla Leigh Pittman Rodgers: Aaron Rodgers' MotherA Spotlight on Darla Leigh Pittman Rodgers: Aaron Rodgers' Mother
A Spotlight on Darla Leigh Pittman Rodgers: Aaron Rodgers' Mother
 
Behind the Scenes The Life of Enzo Zelocchi, a Hollywood Film Producer.pdf
Behind the Scenes The Life of Enzo Zelocchi, a Hollywood Film Producer.pdfBehind the Scenes The Life of Enzo Zelocchi, a Hollywood Film Producer.pdf
Behind the Scenes The Life of Enzo Zelocchi, a Hollywood Film Producer.pdf
 
NO1 Certified kala ilam Expert In Peshwar Kala Jadu Specialist In Peshwar Kal...
NO1 Certified kala ilam Expert In Peshwar Kala Jadu Specialist In Peshwar Kal...NO1 Certified kala ilam Expert In Peshwar Kala Jadu Specialist In Peshwar Kal...
NO1 Certified kala ilam Expert In Peshwar Kala Jadu Specialist In Peshwar Kal...
 
Biswanath Byam Samiti Open Quiz 2022 by Qui9 Grand Finale
Biswanath Byam Samiti Open Quiz 2022 by Qui9 Grand FinaleBiswanath Byam Samiti Open Quiz 2022 by Qui9 Grand Finale
Biswanath Byam Samiti Open Quiz 2022 by Qui9 Grand Finale
 
ECOLUXE pre-ESPYS Ultimate Sports Lounge 2024
ECOLUXE pre-ESPYS Ultimate Sports Lounge 2024ECOLUXE pre-ESPYS Ultimate Sports Lounge 2024
ECOLUXE pre-ESPYS Ultimate Sports Lounge 2024
 
Aesthetic Design Inspiration by Slidesgo.pptx
Aesthetic Design Inspiration by Slidesgo.pptxAesthetic Design Inspiration by Slidesgo.pptx
Aesthetic Design Inspiration by Slidesgo.pptx
 
Moveable Feast_Travel-Lifestyle-Culture Quiz.pptx
Moveable Feast_Travel-Lifestyle-Culture Quiz.pptxMoveable Feast_Travel-Lifestyle-Culture Quiz.pptx
Moveable Feast_Travel-Lifestyle-Culture Quiz.pptx
 
Statement Of Intent - - Copy.documentfile
Statement Of Intent - - Copy.documentfileStatement Of Intent - - Copy.documentfile
Statement Of Intent - - Copy.documentfile
 
NO1 Certified Black magic specialist,Expert in Pakistan Amil Baba kala ilam E...
NO1 Certified Black magic specialist,Expert in Pakistan Amil Baba kala ilam E...NO1 Certified Black magic specialist,Expert in Pakistan Amil Baba kala ilam E...
NO1 Certified Black magic specialist,Expert in Pakistan Amil Baba kala ilam E...
 
Taken Pilot Episode Story pitch Document
Taken Pilot Episode Story pitch DocumentTaken Pilot Episode Story pitch Document
Taken Pilot Episode Story pitch Document
 

En vedette

2024 State of Marketing Report – by Hubspot
2024 State of Marketing Report – by Hubspot2024 State of Marketing Report – by Hubspot
2024 State of Marketing Report – by HubspotMarius Sescu
 
Everything You Need To Know About ChatGPT
Everything You Need To Know About ChatGPTEverything You Need To Know About ChatGPT
Everything You Need To Know About ChatGPTExpeed Software
 
Product Design Trends in 2024 | Teenage Engineerings
Product Design Trends in 2024 | Teenage EngineeringsProduct Design Trends in 2024 | Teenage Engineerings
Product Design Trends in 2024 | Teenage EngineeringsPixeldarts
 
How Race, Age and Gender Shape Attitudes Towards Mental Health
How Race, Age and Gender Shape Attitudes Towards Mental HealthHow Race, Age and Gender Shape Attitudes Towards Mental Health
How Race, Age and Gender Shape Attitudes Towards Mental HealthThinkNow
 
AI Trends in Creative Operations 2024 by Artwork Flow.pdf
AI Trends in Creative Operations 2024 by Artwork Flow.pdfAI Trends in Creative Operations 2024 by Artwork Flow.pdf
AI Trends in Creative Operations 2024 by Artwork Flow.pdfmarketingartwork
 
PEPSICO Presentation to CAGNY Conference Feb 2024
PEPSICO Presentation to CAGNY Conference Feb 2024PEPSICO Presentation to CAGNY Conference Feb 2024
PEPSICO Presentation to CAGNY Conference Feb 2024Neil Kimberley
 
Content Methodology: A Best Practices Report (Webinar)
Content Methodology: A Best Practices Report (Webinar)Content Methodology: A Best Practices Report (Webinar)
Content Methodology: A Best Practices Report (Webinar)contently
 
How to Prepare For a Successful Job Search for 2024
How to Prepare For a Successful Job Search for 2024How to Prepare For a Successful Job Search for 2024
How to Prepare For a Successful Job Search for 2024Albert Qian
 
Social Media Marketing Trends 2024 // The Global Indie Insights
Social Media Marketing Trends 2024 // The Global Indie InsightsSocial Media Marketing Trends 2024 // The Global Indie Insights
Social Media Marketing Trends 2024 // The Global Indie InsightsKurio // The Social Media Age(ncy)
 
Trends In Paid Search: Navigating The Digital Landscape In 2024
Trends In Paid Search: Navigating The Digital Landscape In 2024Trends In Paid Search: Navigating The Digital Landscape In 2024
Trends In Paid Search: Navigating The Digital Landscape In 2024Search Engine Journal
 
5 Public speaking tips from TED - Visualized summary
5 Public speaking tips from TED - Visualized summary5 Public speaking tips from TED - Visualized summary
5 Public speaking tips from TED - Visualized summarySpeakerHub
 
ChatGPT and the Future of Work - Clark Boyd
ChatGPT and the Future of Work - Clark Boyd ChatGPT and the Future of Work - Clark Boyd
ChatGPT and the Future of Work - Clark Boyd Clark Boyd
 
Getting into the tech field. what next
Getting into the tech field. what next Getting into the tech field. what next
Getting into the tech field. what next Tessa Mero
 
Google's Just Not That Into You: Understanding Core Updates & Search Intent
Google's Just Not That Into You: Understanding Core Updates & Search IntentGoogle's Just Not That Into You: Understanding Core Updates & Search Intent
Google's Just Not That Into You: Understanding Core Updates & Search IntentLily Ray
 
Time Management & Productivity - Best Practices
Time Management & Productivity -  Best PracticesTime Management & Productivity -  Best Practices
Time Management & Productivity - Best PracticesVit Horky
 
The six step guide to practical project management
The six step guide to practical project managementThe six step guide to practical project management
The six step guide to practical project managementMindGenius
 
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...RachelPearson36
 

En vedette (20)

2024 State of Marketing Report – by Hubspot
2024 State of Marketing Report – by Hubspot2024 State of Marketing Report – by Hubspot
2024 State of Marketing Report – by Hubspot
 
Everything You Need To Know About ChatGPT
Everything You Need To Know About ChatGPTEverything You Need To Know About ChatGPT
Everything You Need To Know About ChatGPT
 
Product Design Trends in 2024 | Teenage Engineerings
Product Design Trends in 2024 | Teenage EngineeringsProduct Design Trends in 2024 | Teenage Engineerings
Product Design Trends in 2024 | Teenage Engineerings
 
How Race, Age and Gender Shape Attitudes Towards Mental Health
How Race, Age and Gender Shape Attitudes Towards Mental HealthHow Race, Age and Gender Shape Attitudes Towards Mental Health
How Race, Age and Gender Shape Attitudes Towards Mental Health
 
AI Trends in Creative Operations 2024 by Artwork Flow.pdf
AI Trends in Creative Operations 2024 by Artwork Flow.pdfAI Trends in Creative Operations 2024 by Artwork Flow.pdf
AI Trends in Creative Operations 2024 by Artwork Flow.pdf
 
Skeleton Culture Code
Skeleton Culture CodeSkeleton Culture Code
Skeleton Culture Code
 
PEPSICO Presentation to CAGNY Conference Feb 2024
PEPSICO Presentation to CAGNY Conference Feb 2024PEPSICO Presentation to CAGNY Conference Feb 2024
PEPSICO Presentation to CAGNY Conference Feb 2024
 
Content Methodology: A Best Practices Report (Webinar)
Content Methodology: A Best Practices Report (Webinar)Content Methodology: A Best Practices Report (Webinar)
Content Methodology: A Best Practices Report (Webinar)
 
How to Prepare For a Successful Job Search for 2024
How to Prepare For a Successful Job Search for 2024How to Prepare For a Successful Job Search for 2024
How to Prepare For a Successful Job Search for 2024
 
Social Media Marketing Trends 2024 // The Global Indie Insights
Social Media Marketing Trends 2024 // The Global Indie InsightsSocial Media Marketing Trends 2024 // The Global Indie Insights
Social Media Marketing Trends 2024 // The Global Indie Insights
 
Trends In Paid Search: Navigating The Digital Landscape In 2024
Trends In Paid Search: Navigating The Digital Landscape In 2024Trends In Paid Search: Navigating The Digital Landscape In 2024
Trends In Paid Search: Navigating The Digital Landscape In 2024
 
5 Public speaking tips from TED - Visualized summary
5 Public speaking tips from TED - Visualized summary5 Public speaking tips from TED - Visualized summary
5 Public speaking tips from TED - Visualized summary
 
ChatGPT and the Future of Work - Clark Boyd
ChatGPT and the Future of Work - Clark Boyd ChatGPT and the Future of Work - Clark Boyd
ChatGPT and the Future of Work - Clark Boyd
 
Getting into the tech field. what next
Getting into the tech field. what next Getting into the tech field. what next
Getting into the tech field. what next
 
Google's Just Not That Into You: Understanding Core Updates & Search Intent
Google's Just Not That Into You: Understanding Core Updates & Search IntentGoogle's Just Not That Into You: Understanding Core Updates & Search Intent
Google's Just Not That Into You: Understanding Core Updates & Search Intent
 
How to have difficult conversations
How to have difficult conversations How to have difficult conversations
How to have difficult conversations
 
Introduction to Data Science
Introduction to Data ScienceIntroduction to Data Science
Introduction to Data Science
 
Time Management & Productivity - Best Practices
Time Management & Productivity -  Best PracticesTime Management & Productivity -  Best Practices
Time Management & Productivity - Best Practices
 
The six step guide to practical project management
The six step guide to practical project managementThe six step guide to practical project management
The six step guide to practical project management
 
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...
 

Minimum message length, simplicity, and truth

  • 1. The promise The puzzles The prospects Minimum message length as a truth-conducive simplicity measure Stephen Petersen steve@stevepetersen.net Department of Philosophy Niagara University Formal Epistemology Workshop 2 June 2007 Steve Petersen MML and simplicity
  • 2. The promise The puzzles The prospects “ How easy it is for the gullible to clutch at the hope that somehow, deep in the cabalistic mysteries of information, computation, G¨del, Escher, and Bach, there really is an o occult connection between simplicity and reality that will direct us unswervingly to the Truth; that prior probabilities constructed to favor computationally compressible strings really are informative and that learning can be defined as data-compression. After all, aren’t these things constructed out of “information”? I know whereof I speak—I have met these glassy-eyed wretches (full professors, even) and they are beyond salvation. ” —Kevin Kelly Steve Petersen MML and simplicity
  • 3. The promise The puzzles The prospects Introduction The problem area philosophy of science epistemology cognitive science Prospects for the Minimum Message Length (MML) algorithm A purported truth-conducive simplicity measure Steve Petersen MML and simplicity
  • 4. The promise The puzzles The prospects Outline MML: the promise 1 Background MML & Bayes MML & Kolmogorov MML: the puzzles 2 MML over Bayes UTMs as universal priors “Information” The data compression analogy MML: the prospects 3 MML over Bayes UTMs as universal priors “Information”, data compression, and the nature of abduction Steve Petersen MML and simplicity
  • 5. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov Shannon information: the basic idea Uncertainty about data source ≈ variety of possible results More possibilities ≈ more uncertainty Entropy as uncertainty-measure Information as reduction in entropy (uncertainty) Steve Petersen MML and simplicity
  • 6. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov Shannon information: an example How many yes/no questions needed to find the king? 6 divisions of possibility space in half 6 (1) = 1 2 64 1 log 1 = log2 64 = 6 64 2 Measure in bits Efficient binary encoding uses 6 digits Steve Petersen MML and simplicity
  • 7. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov Shannon information: information and entropy Risky approach: e.g., “is it on square f7?” If yes, we gain 6 bits of information If no, we gain only about .023 bits The ideal strategy maximizes information gained with either answer Steve Petersen MML and simplicity
  • 8. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov Shannon information: subjectivity 1 Suppose odds were the king is on its original square 2 Clever first question: “is it on its home square?” Now on average 3.99 yes/no questions needed The entropy now 3.99 bits New digital encoding wise, too Lower entropy reflects receiver’s lower uncertainty about source A subjective measure Steve Petersen MML and simplicity
  • 9. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov Shannon information: definitions 1 Shannon information: h(x) = log2 P(x) Shannon message length: Lens (x) = h(x) Shannon entropy: H(X ) = ∑x P(x)h(x) Steve Petersen MML and simplicity
  • 10. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov Kolmogorov complexity in one slide Universal Turing Machine U, target string s “Description” dU (s): the shortest input to U that returns s Kolmogorov complexity KU (s) = |dU (s)| Detectable patterns (repeated sequences, the digits of π in binary . . . ) → shorter programs Interesting measure of randomness Lens (s) and K (s) closely related Steve Petersen MML and simplicity
  • 11. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov MML: The basic idea Bayes: most probable hypothesis given the data Shannon: more probable claims allow shorter messages MML: find most probable hypothesis by shortest message ≈ find truth via simplicity Steve Petersen MML and simplicity
  • 12. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov MML is truth-conducive: the “proof” We have truth-related reason to pick P(hi , e) argmax P(hi |e) = argmax P(e) i i = argmax P(hi , e) i 1 = argmin log2 P(hi , e) i = argmin Lens (hi , e) i 1 Figure: log2 P(x) Steve Petersen MML and simplicity
  • 13. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov Simplicity and data-fit Epistemic reason to pick shortest statement of hypothesis with data Big rabbit, small hat Note especially: Lens (hi , e) = − log2 P(hi , e) = − log2 P(hi )P(e|hi ) = − log2 P(hi ) + − log2 P(e|hi ) = Lens (hi ) + Lens (e|hi ) Balance of simplicity and data-fit! Steve Petersen MML and simplicity
  • 14. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov The MML explanation Table: Wallace’s “explanation” Lens (h) Lens (e|h) “assertion” “details” curve data error pattern noise regularities exceptions compressor compressed file Steve Petersen MML and simplicity
  • 15. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov Explaining to a UTM The eyes get still glassier . . . Assume the Shannon-encoding scheme h is computable Program a Universal Turing Machine U to compute it with hU U determines an implicit prior distribution HU for such programs Lens (h|HU ) ≈ KU (hU ) = |hU | Lens (e|h, HU ) ≈ KU|h (e) Steve Petersen MML and simplicity
  • 16. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov UTMs as uninformative Bayesian priors Explanation lengths will depend on choice of UTM But this difference is bounded above by a constant “The” UTM (up to a constant) as a maximally uninformative prior A step toward a non-subjective information? Steve Petersen MML and simplicity
  • 17. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov Algorithmic science Wallace suggests we insist U|h also be a UTM An “educated” UTM Its implicit prior favors h Can be further educated with any (computable) hypothesis “Normal” and “revolutionary” algorithmic science Steve Petersen MML and simplicity
  • 18. The promise Background The puzzles MML & Bayes The prospects MML & Kolmogorov The robot scientist Steve Petersen MML and simplicity
  • 19. MML over Bayes The promise UTMs as priors The puzzles “Information” The prospects The data compression analogy How is MML any better than Bayesian inference? You might reasonably ask: “ Why do we need MML? Why don’t we just pick the hypothesis with the highest Bayesian posterior probability? ” Steve Petersen MML and simplicity
  • 20. MML over Bayes The promise UTMs as priors The puzzles “Information” The prospects The data compression analogy UTMs as universal priors The “up to constant” part is rarely negligible Suppose UTM2 emulates UTM1 with just a 272-bit program P(h|UTM2 ) = 2−272 · P(h|UTM1 ) Steve Petersen MML and simplicity
  • 21. MML over Bayes The promise UTMs as priors The puzzles “Information” The prospects The data compression analogy The notion of “information” Kevin Kelly is (rightly) concerned about abuses of “information” Ambiguity: Shannon-information vs. learning theory information Compare the ambiguity in a word’s “representing” a thing, and a thinker’s “representing” a thing The former is “derived”, the latter “original” I suspect this is the same ambiguity Steve Petersen MML and simplicity
  • 22. MML over Bayes The promise UTMs as priors The puzzles “Information” The prospects The data compression analogy The null hypothesis One might get the impression that Lens (h) + Lens (e|h) should be shorter than Lens (e) Wallace encourages this with his translation to data compression Also with his discussion of the null hypothesis: “ We will require the length of the explanation message which states and uses an acceptable theory to be shorter than any message restating the data using only the implications of prior premises. ” But this is never the case! Steve Petersen MML and simplicity
  • 23. MML over Bayes The promise UTMs as priors The puzzles “Information” The prospects The data compression analogy Data compression is impossible? Suppose for contradiction Lens (h) + Lens (e|h) < Lens (e) − log P(h) + − log P(e|h) < − log P(e) − log P(h)P(e|h) < − log P(e) − log P(h, e) < − log P(e) P(h, e) > P(e) Straightforwardly contradicts probability axioms There’s some important disanalogy with data compression Steve Petersen MML and simplicity
  • 24. MML over Bayes The promise UTMs as priors The puzzles “Information” The prospects The data compression analogy Too much information Elsewhere Wallace acknowledges this The explanation will exceed encoded data length by Lens (h|e) Still elsewhere, Wallace takes this as a virtue of MML Lens (e, h) > Lens (e) because MML is abductive Why send this extra information, given the cost? Steve Petersen MML and simplicity
  • 25. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction The advantage of MML over Bayes For discrete H , MML = Bayesian inference MML has the advantage for continuous-valued H Bayes: a posterior density, sensitive to parametrization of priors MML finds the right discretization H ∗ And the right h ∈ H ∗ Steve Petersen MML and simplicity
  • 26. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction A classic example Infer coin bias from a series of flips. Table: Potential explanation lenghts (in nits) for 20 heads in 100 h∈H ∗ |H ∗| Lens (h) Lens (e|h) Total length 0 0 51.900 51.900 0 / .20 100 4.615 50.040 54.655 .205 99 3.922 50.045 53.970 .50 1 0 69.315 69.315 .25 10 1.907 50.161 52.068 Steve Petersen MML and simplicity
  • 27. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction Bayes, MML, and Kolmogorov MML’s real strength over Bayesian inference MML also gives approximations for Kolmogorov complexity A handy bridge between the two Neither MML nor Kolmogorov complexity are computable Steve Petersen MML and simplicity
  • 28. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction The simplest UTM “ . . . the complexity of a TM is monotone increasing with its number of states. An overly simple TM, say one with only one or two states, cannot be universal. There must be, and is, a simplest UTM, or a small set of simplest UTMs. Adoption of such a UTM as receiver can reasonably be regarded as expressing no expectation about the theory or estimate to be inferred, save that it will be computable . . . Adoption of any UTM (or TM) with more states seems necessary to assume something extra, i.e., to adopt a “less ignorant” prior. We therefore suggest that the only prior expressing total ignorance is that implied by a simplest UTM. ” Steve Petersen MML and simplicity
  • 29. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction “The” simplest UTM Will still depend on the way of specifying UTMs Wallace says the “simplest” in any such should do Is that right? Would an educated “simplest” UTM use “green” and not “grue”? Steve Petersen MML and simplicity
  • 30. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction The simplest UTM? Wallace may be wrong that a UTM “with only one or two states cannot be universal” Wolfram thinks this “two state, three color” one might be: $25,000 prize! (http://wolframprize.org) Steve Petersen MML and simplicity
  • 31. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction Naturalized intentionality “Original information” as “original intentionality” Naturalized theories of intentionality Millikan-Dretske When it’s a function of a creature to have an internal element covary with aspects of external circumstances, that element can be representational if such covariance is supposed to help satisfy a need of the creature. Steve Petersen MML and simplicity
  • 32. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction Robot scientist representing Millikan: the “consumer-side” makes some symbol representational Our robot scientist could gain “original information” via MML (Assuming that inferring hypotheses meant to covary with a data source can help serve its needs) The priors in these needs? Steve Petersen MML and simplicity
  • 33. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction Solomonoff’s specter Again, why abduction? Could carry forward the entire posterior distribution over H Solomonoff’s “algorithmic probability” predictive strategy Wallace: Solomonoff will predict better, but “ MML attempts to mimic the discovery of natural laws, whereas Solomonoff wants to predict what will happen next with no explicit concern for understanding why. ” Solomonoff is deductive, MML ampliative Steve Petersen MML and simplicity
  • 34. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction Why abduction? Again, why abduction? Especially given cost in message length and predictive power? Marcus Hutter’s “universal AI” based on Solomonoff Hutter: artificial intelligence ≈ data compression 50,000 prize! (http://prize.hutter1.net) Steve Petersen MML and simplicity
  • 35. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction Abduction and AI Hunch: AI as the problem of lossy compression Glean the important parts, toss the rest “Important” parts agent-relative Compare consumer-side intentionality Deductive methods remain badly intractable—maybe not coincidentally MML might help with these (related?) problems Abduction as “mere heuristics” of intelligent creatures Steve Petersen MML and simplicity
  • 36. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction Summary MML is not A source of non-subjective information A clear, clean solution to the problem of the priors A miraculous oracle of truth from simplicity MML is A decent algorithm for Bayesian inference over continuous hypothesis spaces A possible step in gaining “real” information A nice bridge of Bayes and Kolmogorov Part of an intriguing model for scientific progress Some truth-related vindication of data compression (a form of simplicity) Steve Petersen MML and simplicity
  • 37. The promise MML over Bayes The puzzles UTMs as priors The prospects The nature of abduction Thanks Thanks to: Dennis Whitcomb, Rutgers Shahed Sharif, Duke Ken Regan, SUNY Buffalo Alex Bertland, Niagara University Niagara University Branden Fitelson and FEW You Steve Petersen MML and simplicity