Within psychology, neuroscience and artificial intelligence, there has been increasing interest in the proposal that the brain builds probabilistic models of sensory and linguistic input: that is, to infer a probabilistic model from a sample. The practical problems of such inference are substantial: the brain has limited data and restricted computational resources. But there is a more fundamental question: is the problem of inferring a probabilistic model from a sample possible even in principle? We explore this question and find some surprisingly positive and general results. First, for a broad class of probability distributions characterized by computability restrictions, we specify a learning algorithm that will almost surely identify a probability distribution in the limit given a finite i.i.d. sample of sufficient but unknown length. This is similarly shown to hold for sequences generated by a broad class of Markov chains, subject to computability assumptions. The technical tool is the strong law of large numbers. Second, for a large class of dependent sequences, we specify an algorithm which identifies in the limit a computable measure for which the sequence is typical, in the sense of Martin-Löf (there may be more than one such measure). The technical tool is the theory of Kolmogorov complexity. We analyze the associated predictions in both cases. We also briefly consider special cases, including language learning, and wider theoretical implications for psychology.

Additional Metadata
Keywords Bayesian brain, identification, Computable measure, Computable probability, Kolmogorov complexity, Learning, Markov chain, Martin-Löf randomness, Strong law of large numbers, Typicality
Persistent URL dx.doi.org/10.1016/j.jmp.2016.11.004
Journal Journal of Mathematical Psychology
Citation
Vitányi, P.M.B, & Chater, N. (2017). Identification of probabilities. Journal of Mathematical Psychology, 76, 13–24. doi:10.1016/j.jmp.2016.11.004