Simplicity Theory
   HOME
*





Simplicity Theory
Simplicity theory is a cognitive theory that seeks to explain the attractiveness of situations or events to human minds. It is based on work done by scientists like behavioural scientist Nick Chater, computer scientist Paul Vitanyi, psychologist Jacob Feldman, and artificial intelligence researchers Jean-Louis Dessalles Dessalles, J.-L. (2013)"Algorithmic simplicity and relevance" In D. L. Dowe (Ed.), Algorithmic probability and friends - LNAI 7070, 119-130. Berlin, D: Springer Verlag. and Jürgen Schmidhuber. It claims that interesting situations appear simpler than expected to the observer. Overview Technically, simplicity corresponds in a drop in Kolmogorov complexity, which means that, for an observer, the shortest description of the situation is shorter than anticipated. For instance, the description of a consecutive lottery draw, such as 22-23-24-25-26-27, is significantly shorter than a typical one, such as 12-22-27-37-38-42. The former requires only one instantiation ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Nick Chater
Nick Chater is Professor of Behavioural Science at Warwick Business School, who works on rationality and language using a range of theoretical and experimental approaches. Education Chater read Psychology at Cambridge University. He first worked at Warwick University in 1996. Career Chater is head of WBS's Behavioural Science group, which is the largest of its kind in Europe. Chater presents the massive open online course ''The Mind Is Flat''. Chater is a member of the UK Committee on Climate Change. He was an advisor to the UK government's Behavioural Insights Team. He is a Fellow of the Cognitive Science Society and the British Academy The British Academy is the United Kingdom's national academy for the humanities and the social sciences. It was established in 1902 and received its royal charter in the same year. It is now a fellowship of more than 1,000 leading scholars spa .... Chater was scientist-in-residence on eight seasons of the Radio 4 series ''The Human ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Paul Vitanyi
Paul may refer to: *Paul (given name), a given name (includes a list of people with that name) *Paul (surname), a list of people People Christianity * Paul the Apostle (AD c.5–c.64/65), also known as Saul of Tarsus or Saint Paul, early Christian missionary and writer *Pope Paul (other), multiple Popes of the Roman Catholic Church *Saint Paul (other), multiple other people and locations named "Saint Paul" Roman and Byzantine empire *Lucius Aemilius Paullus Macedonicus (c. 229 BC – 160 BC), Roman general *Julius Paulus Prudentissimus (), Roman jurist *Paulus Catena (died 362), Roman notary *Paulus Alexandrinus (4th century), Hellenistic astrologer *Paul of Aegina or Paulus Aegineta (625–690), Greek surgeon Royals * Paul I of Russia (1754–1801), Tsar of Russia *Paul of Greece (1901–1964), King of Greece Other people * Paul the Deacon or Paulus Diaconus (c. 720 – c. 799), Italian Benedictine monk *Paul (father of Maurice), the father of Maurice, By ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Artificial Intelligence
Artificial intelligence (AI) is intelligence—perceiving, synthesizing, and inferring information—demonstrated by machines, as opposed to intelligence displayed by animals and humans. Example tasks in which this is done include speech recognition, computer vision, translation between (natural) languages, as well as other mappings of inputs. The ''Oxford English Dictionary'' of Oxford University Press defines artificial intelligence as: the theory and development of computer systems able to perform tasks that normally require human intelligence, such as visual perception, speech recognition, decision-making, and translation between languages. AI applications include advanced web search engines (e.g., Google), recommendation systems (used by YouTube, Amazon and Netflix), understanding human speech (such as Siri and Alexa), self-driving cars (e.g., Tesla), automated decision-making and competing at the highest level in strategic game systems (such as chess and Go). ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Jean-Louis Dessalles
Jean-Louis Dessalles (born in 1956 in Périgueux) is a French computer scientist and researcher in artificial intelligence and cognitive science, professor à Télécom Paris (Paris). He is best known for his contributions to the Simplicity theory and for his original theory about a possible political origin of language. Biography Dessalles was born in Périgueux in Southwestern France. He graduated from École Polytechnique in 1979 (promotion X76) and from Télécom ParisTech in 1981. He is currently Professor of Computer Science at Télécom Paris, which is part of the University of Paris-Saclay. Research ''Jean-Louis Dessalles'' focuses on the quest for fundamental principles underlying the language faculty and its biological origins. His contribution to Simplicity theory was to show that complexity ''drop'' predicts narrative interest. He also designed a concise model of argumentative relevance. On the issue of human language evolutionary origins, he found that Cos ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  




Jürgen Schmidhuber
Jürgen Schmidhuber (born 17 January 1963) is a German computer scientist most noted for his work in the field of artificial intelligence, deep learning and artificial neural networks. He is a co-director of the Dalle Molle Institute for Artificial Intelligence Research in Lugano, in Ticino in southern Switzerland. Following Google Scholar, from 2016 to 2021 he has received more than 100,000 scientific citations. He has been referred to as "father of modern AI," "father of AI," "dad of mature AI," "Papa" of famous AI products, "Godfather," and "father of deep learning." (Schmidhuber himself, however, has called Alexey Grigorevich Ivakhnenko the "father of deep learning.") Schmidhuber completed his undergraduate (1987) and PhD (1991) studies at the Technical University of Munich in Munich, Germany. His PhD advisors were Wilfried Brauer and Klaus Schulten. He taught there from 2004 until 2009 when he became a professor of artificial intelligence at the Università della Sviz ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Kolmogorov Complexity
In algorithmic information theory (a subfield of computer science and mathematics), the Kolmogorov complexity of an object, such as a piece of text, is the length of a shortest computer program (in a predetermined programming language) that produces the object as output. It is a measure of the computational resources needed to specify the object, and is also known as algorithmic complexity, Solomonoff–Kolmogorov–Chaitin complexity, program-size complexity, descriptive complexity, or algorithmic entropy. It is named after Andrey Kolmogorov, who first published on the subject in 1963 and is a generalization of classical information theory. The notion of Kolmogorov complexity can be used to state and prove impossibility results akin to Cantor's diagonal argument, Gödel's incompleteness theorem, and Turing's halting problem. In particular, no program ''P'' computing a lower bound for each text's Kolmogorov complexity can return a value essentially larger than ''P'''s own leng ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Algorithmic Probability
In algorithmic information theory, algorithmic probability, also known as Solomonoff probability, is a mathematical method of assigning a prior probability to a given observation. It was invented by Ray Solomonoff in the 1960s. It is used in inductive inference theory and analyses of algorithms. In his general theory of inductive inference, Solomonoff uses the method together with Bayes' rule to obtain probabilities of prediction for an algorithm's future outputs. In the mathematical formalism used, the observations have the form of finite binary strings viewed as outputs of Turing machines, and the universal prior is a probability distribution over the set of finite binary strings calculated from a probability distribution over programs (that is, inputs to a universal Turing machine). The prior is universal in the Turing-computability sense, i.e. no string has zero probability. It is not computable, but it can be approximated. Overview Algorithmic probability is the main ingre ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Subjective Probability
Bayesian probability is an interpretation of the concept of probability, in which, instead of frequency or propensity of some phenomenon, probability is interpreted as reasonable expectation representing a state of knowledge or as quantification of a personal belief. The Bayesian interpretation of probability can be seen as an extension of propositional logic that enables reasoning with hypotheses; that is, with propositions whose truth or falsity is unknown. In the Bayesian view, a probability is assigned to a hypothesis, whereas under frequentist inference, a hypothesis is typically tested without being assigned a probability. Bayesian probability belongs to the category of evidential probabilities; to evaluate the probability of a hypothesis, the Bayesian probabilist specifies a prior probability. This, in turn, is then updated to a posterior probability in the light of new, relevant data (evidence). The Bayesian interpretation provides a standard set of procedures and formula ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Subjective Probability
Bayesian probability is an interpretation of the concept of probability, in which, instead of frequency or propensity of some phenomenon, probability is interpreted as reasonable expectation representing a state of knowledge or as quantification of a personal belief. The Bayesian interpretation of probability can be seen as an extension of propositional logic that enables reasoning with hypotheses; that is, with propositions whose truth or falsity is unknown. In the Bayesian view, a probability is assigned to a hypothesis, whereas under frequentist inference, a hypothesis is typically tested without being assigned a probability. Bayesian probability belongs to the category of evidential probabilities; to evaluate the probability of a hypothesis, the Bayesian probabilist specifies a prior probability. This, in turn, is then updated to a posterior probability in the light of new, relevant data (evidence). The Bayesian interpretation provides a standard set of procedures and formula ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]