Friday, November 30, 2018

The Communist Party of Britain



I'm quite nostalgic about my early twenties in the International Marxist Group, the British Section of the Fourth International. The cause was exciting and compelling, the organisation provided a normative framework for day-to-day life, there were great parties and discos, you had a support network wherever you went .. and as there were sections in most countries revolutionary tourism was a super-benefit (stay with the local comrades in the south of France - as I did, twice).

Wouldn't it be nice to replicate that sense of belonging? Clare has the Catholic Church, another internationalist, mass-participation organisation, but it's not for me. The Fourth International too is a lost cause. In the UK it's now just another SJW group with somewhat better economics.

But the Communist Party?

I had to check, I thought it had disbanded.

Not so. The Communist Party of Britain is alive and publishes the Morning Star. More than that, it's plugged into the Corbyn project.

It's quite well-connected. When it organised the Communist University at Ruskin House in November 2006 speakers included Labour MP John McDonnell, RMT general secretary Bob Crow and CND chair Kate Hudson. Corbyn reads and writes for the Morning Star.

Admittedly its membership is small and declining. In 2017 its adherents numbered 734 (a figure even the IMG attained once). In elections its votes have been derisory.

---

Still, who needs popularity if you're right. There's a branch about twenty miles away. Perhaps I should check what it stands for?

From the Party website FAQs (my emphasis throughout):
"What about the crimes committed by Stalin?

The crimes committed during the Stalin period cannot be ignored. But the first attempts to build a socialist society took place in a semi-feudal society facing the hostile forces of imperialism. A bureaucratic-command system of economic and political rule became entrenched.  The Communist Party of the Soviet Union and the trade unions became integrated into the apparatus of the state, eroding working class and popular democracy.

In the late 1930s in particular, severe violations of socialist democracy and law occurred as large numbers of innocent people were imprisoned and executed. Lessons have to be learnt from both the achievements and the failures of this period."
"Eroding"? I believe this is called a (tepid) stalinist apologia. Mistakes were made and lessons must be learned. What and what?
"Doesn't communism go against human nature?

The idea that human nature and communism don't mix is based on the concept that humans are innately individualistic and selfish. In truth, human behaviour is related to the society in which they develop. Capitalism actively encourages these behavioural traits to develop and become normal in society. Under socialism the conditions for these traits to be advantageous are removed and instead are replaced by those that promote mutual cooperation and respect for other members of society."
Good luck with that.
"What is the CPBs' position on the EU?

We're a party of internationalists so we're committed to supporting the workers' and peoples' of Europe in their struggles. Ultimately we consider the cultural, historic and other differences between workers in different countries to be less significant than with the differences between Britain's workers and members of the British capitalist class. However the EU is an institution that is fundamentally opposed to working class interests. The EU exists to promote the interest of monopoly capital and big business, particularly German and French.  ...

We want to see a Europe of democratic states that value public services, guarantees the rights of workers and puts the interests of ordinary people above those of big business. This is impossible while we remain within the EU given its anti-democratic structures. We need to campaign for a ‘left exit’ from the EU, based on socialist policies that respect the rights of other European countries."
Why isn't this Labour's policy? 

[I know why].

---

The Party's programme is called Britain's Road to Socialism.
"Britain's Road to Socialism is the programme of the Communist Party of Britain, and is adhered to by the Young Communist League and the editors of the Morning Star newspaper.

It proposes that socialism can be achieved in Britain by the working class leading the other classes in a popular democratic anti-monopoly alliance against monopoly capital, and implementing a left-wing programme of socialist construction."
Ninety eight years of existence as a Communist Party and these folk have learned nothing. The muscular old-time CPGB reimagined as a senior citizen tribute band.

I rather fear I shall have to look elsewhere for my concierge organisation of choice.

Wednesday, November 28, 2018

Forensic psychology: struggling to become a science

Amazon link

I received a free copy of this to review through the Amazon Vine programme. It's a lengthy textbook and I won't be finished anytime soon. But reading the introduction and initial chapters is enough to discern a profession transitioning through intellectual crisis.

Here are my notes so far.

---

"We do have a scientific paradigm for psychology. We know that genes and environment construct brains, brains produce behaviour. Individual differences in genome plus environmental divergences result in non-identical brains, differences which have a measurable effect on neuroanatomy .. and thence on behaviour.

From behavioural genetics we know that the heritability of psychological traits is generally more than 50%. Nurturing differences are often really the consequence of genetic correlations: disturbed parents tend to have disturbed children. Heritability in action. Non-shared environmental effects (eg not family) typically have random and transient effects on psychological traits such as personality and intelligence. Real abuse and damage always excepted.

Psychology sort of knows this and sort of doesn't. There are researchers who know what a GWAS is, who know their way round a brain scan. And then there is the long legacy of pure nurturism:  behaviourism, psychoanalysis, attachment theory, sociomoral reasoning, social information-processing theory and so on.

These one-sided plausible-sounding abstractions of commonplace observations have sadly produced neither real insight, reproducible results nor consistently-effective interventions. In many cases large scale experiments using twin studies and GWAS have shown that their conclusions are simply false. Perhaps though this will change, now we're beginning to really understand the etiology of normal and abnormal psychology. Robert Plomin's "Blueprint" is a readable review covering all these points.

All this has additional force in the case of forensic psychology. But who would wish to be in the position of the textbook editor in 2018, forced to straddle competing and incompatible paradigms while maintaining that forensic psychology is a mature discipline on firm foundations, whose professionals can be relied upon in the investigation process, in court and during the follow-on treatment of prisoners.

Forensic psychology has two main aspects: the legal aspect deals with evidence, witnesses and the courts while the criminological aspect deals with crime and criminals. Our textbook starts by reviewing theories addressing the causes of crime: why do criminals offend and what is the effect on their victims? As mentioned above there are many kinds of theory, and they tell very different kinds of story.

Here is an illustrative quote (p. 58, section 2.2.1 - LCPs are life-course persistent offenders).
“The main factors that encourage offending by the LCPs are cognitive deficits, an under-controlled temperament, hyperactivity, poor parenting, disrupted families, teenage parents, poverty and low socioeconomic status (SES).

Genetic and biological factors, such as a low heart rate, are also important (see Chapter 4).”

Of course all of the so-called main factors on the first list are strongly heritable, and therefore have substantial genetic causation which would be captured by a polygenic score (not mentioned in the copious index).

Amazing to read this in a 2018 textbook."

Full review here.

Tuesday, November 27, 2018

Government as a stationary bandit

Amazon link

I would call this cynical, except it's economics:
“Consider a roving bandit who roams from village to village, looting village residents. The roving bandit will take as much as he can before moving on to his next target. The bandit gets no advantage from leaving anything behind for the people who are plundered. The roving bandit must continually be on the move to find new targets, and resources are wasted on both sides as the bandit uses force to try to plunder the villagers and the villagers use force to try to repel the bandit.

An alternative strategy would be for the bandit to occupy a location permanently, becoming a stationary bandit, periodically taking resources from those in the village.

The incentives are different for the stationary bandit, who would then want to cultivate the productivity of those he plunders, so there is more to plunder in the future.

Because the income of the stationary bandit comes from the productivity of those under him, the stationary bandit also has an incentive to protect them, both from opportunistic individuals in their own village and from outside predators. It is easy to imagine that a productive village would attract the attention of other predators who might be inclined to attack and plunder the village. So, the stationary bandit has an incentive to protect those who provide the stationary bandit’s income.

It is easier to gain compliance from people when they agree to comply than when they are forced, so the stationary bandit has an incentive to get those under him to view his activities as legitimate: to present himself as a benevolent leader rather than a bandit.

If he can convince citizens to recognize him as a king, and can convince them that it is to their advantage to pay him taxes in exchange for the protection he offers, it will be less costly for him to maintain a system in which citizens view the transfer of resources to him as legitimate taxation rather than plunder.

The stationary bandit has an incentive to establish institutions such that citizens view it as advantageous to transfer resources to the ruler in exchange for protection.

The ruler also has an incentive to produce public goods if they enhance the productivity of citizens, because greater productivity means there is more to tax. And, the ruler has an incentive to place constitutional limits on his power to tax, to maintain incentives for citizens to be productive. If the ruler is perceived as a threat to confiscate everything the citizens have, they will not have much of an incentive to accumulate wealth.

Establishing a council of citizens who must approve any tax increases could work to the mutual advantage of both the ruler and the citizens, making citizens more productive and ultimately tendering more tax revenue from their increased productivity.

If a stationary bandit envisions remaining in place for a long time, he will have an incentive to establish institutions that would be very similar to what citizens might agree to from behind a veil of ignorance. Thus, when thinking about optimal constitutional rules, there may not be that much difference between the types of rules people would voluntarily adopt and those which might be imposed by a conquering group.

This public choice approach to constitutional rules offers an interesting contrast to the public goods theory that economists often use to explain government production.

For example, public goods theory suggests that government produces national defense because it is a public good that would be underprovided in the market.

This public choice approach says that government produces national defense because it benefits those in government by protecting the productivity of those who pay taxes to government. Public goods theory depicts government as a benevolent provider of a good that the selfish private sector will underprovide to themselves.

Are people in government really more benevolent than people in the private sector? This public choice approach depicts the production of national defense as a mutually advantageous exchange in which everyone acts to further their own interests, so is more consistent with the assumptions economists typically assume about individual behavior.” (pp. 131-133).
This is not such a new idea. Razib Khan recounted how:
"During the Mongol conquest of Northern China Genghis Khan reputedly wanted to turn the land that had been the heart of the Middle Kingdom into pasture, first by exterminating the whole population.

Part of the motive was to punish the Chinese for resisting his armies, and part of it was to increase his wealth. One of his advisors, Yelu Chucai, a functionary from the Khitai people, dissuaded him from this path through appealing to his selfishness.

Chinese peasants taxed on their surplus would enrich Genghis Khan far more than enlarging his herds. Rather than focus on primary production, Genghis Khan could sit atop a more complex economic system and extract rents."
Public choice theory is a lot better than the soppier humanities, which assume everyone is endlessly caring and benevolent (and should therefore be punished if they are not). But in starting from the fiction that society is simply a uniform set of individual, selfish, rational agents making more-or-less informed utilitarian choices, it abstracts from the historical evolution of class relations, the conditions of social reproduction, and human nature itself.




That's not to say that under certain localised conditions of stabilised atomisation, it can't be quite accurate and predictive, as János Kornai documented.

Monday, November 26, 2018

The relative autonomy of the petty bourgeoisie

Nicos Poulantzas

Sociologists Bradley Campbell and Jason Manning came to people's attention in early 2018 with their book, "The Rise of Victimhood Culture: Microaggressions, Safe Spaces, and the New Culture Wars". Seeking to account for an SJW phenomenon increasingly dominant in the western world, they distinguished between three kinds of culture: Honor, Dignity, and Victimhood.

Campbell and Manning over-abstracted these cultural categories, representing them almost as menu choices, or contingent way-points on the arc of history. This was mostly corrected in Tanner Greer's commentary at "The Scholar's Stage" (which features on the sidebar in the web version of this blog).

Rather than having me summarise his impressive analysis, please review it for yourself: "Honor, Dignity, and Victimhood: A Tour Through Three Centuries of American Political Culture"

Greer's analysis was not perfect.
"The transition from honor to dignity reflected a deeper shift in the basic building blocks of society. Had Western Europeans not shifted away from clan based to citizen based communities, there would not have been any Culture of Dignity. The shift from dignity to victimhood reflects a similar change in the societal distribution of power. ...

The federal government assumes powers traditionally reserved to local and state governments. Local businesses have been pushed out of existence by international conglomerates. The businesses, associations, congregations, and clubs that once made up American society are gone. America has been atomized; her citizens live alone, connected but weakly one to another. Arrayed against each is a set of vast, impersonal bureaucracies that cannot be controlled, only appealed to.

A "Culture of Victimhood" is a perfectly natural response to this shift in the distribution of power. Remember that the central purpose of moral cultures is to help resolve or deter disputes. Dignity cultures provide a moral code to regulate disputes among equals from the same community. They also help individuals in a community--citizens--organize to protect their joint interests.

21st century America has lost this ability to organize and solve problems at the local level. The most effective way to resolve disputes is appeal to the powerful third parties: corporations, the federal government, or the great mass of people weakly connected by social media. The easiest way to earn the sympathy of these powers is to be the unambiguous victim in the dispute.

Victim culture is here to stay."
In fact this 'atomisation' has to be interpreted more dynamically, in class terms. After all, not everyone identifies as a 'victim'.

Peter Turchin has correctly linked neoliberal atomisation with the 'overproduction of elites', the mass production of (social science and humanities) graduates with inflated expectations and entitlements, but for whom no place exists within the real elite, the top 1% or 0.1%. They are angry and febrile and in the mass a cause of political instability. It's not a new phenomenon: some of us remember the late sixties and early seventies.

This (now vastly enlarged) intermediate layer between the traditional working class and the high bourgeoisie is termed by Marxists the petty bourgeoisie. This brought to mind the seminal work of theorist Nicos Poulantzas and I reviewed with interest his chapter 13, "The New Petty Bourgeoisie" in "The Poulantzas Reader" (PDF) .

Poulantzas was an interesting guy and I also reviewed his bio, nodding along, until I came to this:
"He killed himself in 1979 by jumping from the window of a friend's flat in Paris."
He was 43.

There was something about the mental stability of continental Marxists in the late twentieth century. His more famous contemporary, Louis Althusser,  killed his wife Hélène Rytmann the following year and was declared unfit to stand trial due to insanity. He was committed to a psychiatric hospital for three years.

Why are the new petty bourgeoisie so restless again? Because a decade of economic stagnation with the increasing gap between their circumstances and those of the top 1% has left them feeling abandoned. They experience a dislocating lack of prospects and purpose.

The old-time (pre-1980s) working class had a historic cultural identity forged through tenement communities, monolithic employers, trade unionism and social-democratic activism. The modern petty bourgeoisie by contrast is a collection of isolates, expressing their angst through faddish identity-issues.

None of these campaigns ever address the real causes of their dissatisfaction. Nevertheless, their heartfelt herd-activism is always destabilising .. and from the bourgeoisie's point of view, always with a risk of triggering wider unrest. Consequently, the elites attempt to manipulate and co-opt them (eg the moralising campaign against plastic pollution which carefully fails to discuss the true sources of said pollution).

The classic Marxist texts (Lenin, Trotsky) recognised the febrile and changeable nature of the radicalised petty bourgeoisie. Their works discuss methods of creating an alliance with the proletariat in the struggle for socialism.

How times have changed now that the tail wags the dog: the era of Momentum and the death of the dream .. .

Saturday, November 24, 2018

UDI from the proto-empire

Brexit: the fundamental issue is the UK vis-à-vis Germany

Brexit is not hard to understand.

1. The Germans are overt players again

Germany is waking from its decades of political hibernation. Its long atonement for the crimes of Nazism. The years in which it hid pursuit of its national interests behind the EU project, passive-aggressively ceding political leadership to France.

The new emerging Germany is finally ready and willing to be the European hegemon. Ask Greece.

2. The French know their place

The French elite, under Macron, has already capitulated, providing ideological cover to the German project with its narrative of transnationalism and 'ever closer union'. It's rather like the UK relationship with the US although the French have more influence over Germany.

La France profonde is increasingly restive.

3. The British: global financial reach + indifferent regional manufacturing

The UK elite, from its vantage point of a global financial nexus, knows it could never dominate Germany in Europe. But neither does it wish to be dominated. Its European strategy has always been transactional and divisive.

4. A political UDI suits both sides in the EU

A rising Germany is happy that the UK, a game-spoiling rival, is withdrawing from the ring. The French likewise.

The more globalised, financialised UK sees a least bad outcome in maintaining optimised economic access to the European market for goods and services. The British (English? .. London?) elite is therefore broadly happy with the May government's Brexit strategy and is content to spend the next plurality of years fine-tuning it.

Despite the aspirations of UK Remainers (fans of the Macron project), there has been no concerted pressure from the major EU powers (Germany, France) for a second referendum. The 'People's Vote’ has been theatre, reflecting a lack of purchase on the core of UK elite opinion.

5. Labour should tactically prop up the May government

If Labour were smart, they would abstain on the early December parliamentary vote on the May proposals.

This would avoid complicity in the risk of an uncontrolled hard Brexit (which would dismay their Momentum base and be strongly opposed by the UK elite) while exposing May's reliance on Labour to get her own policy through.

The three-way civil war inside the Conservative party would intensify and the Tories would look even less a viable government.

6. Plus ça change ... in the short run

As the UK transitions out of the EU in March 2019, all the cans will have been kicked down the road for another few years.

The May government may well stagger on to the next election (due 2022), spinning its administrative wheels with no consensual Tory vision or roadmap whatsoever - and with Labour salivating and preparing for government.

Friday, November 23, 2018

Understanding Polygenic Scores (PGS)

In the years to come it will be very important to understand the concept of your polygenic score for traits such as height, weight, intelligence, personality and many others. In chapter 12 of his book, "Blueprint: How DNA Makes Us Who We Are" Robert Plomin gives a gentle introduction to the PGS concept which I excerpt here.

---

"Because polygenic scores are the basis for the DNA revolution in psychology, it is essential to understand what they are. A polygenic score is like any composite score that psychologists routinely use to create scales from items, such as those on a personality questionnaire. The goal of a polygenic score is to provide a single genetic index to predict a trait, whether schizophrenia, well-being or intelligence.

To get a concrete understanding of a polygenic score, consider a personality trait like shyness. A questionnaire to assess shyness includes multiple items in order to tap into different facets of shyness. For example, a typical shyness questionnaire will have items about how anxious you are in social situations and how much you avoid these situations for example, going to a party, meeting strangers and speaking up at a meeting. You might be asked to respond using a three-point scale (0 = not at all, 1 = sometimes, 2 = a lot).

A shyness score is created by adding these items, taking care to ‘reverse’ items as needed so that a high score means a high degree of shyness. If our shyness measure had ten items scored 0, 1 and 2, total scores could vary from 0 to 20. Simply adding the items like this treats each item as if it is equally useful, but all items are not equally useful. For this reason, items are often added after they are weighted by some criterion of their usefulness at capturing the construct of shyness.

This is exactly how polygenic scores are created, except that, instead of items on a questionnaire, we add up SNP genotypes. Like the three-point rating scale for shyness, SNP genotypes are scored as 0, 1 or 2, indicating the number of ‘increasing’ alleles, as in the example of the FTO SNP [a polymorphism implicated in weight gain].

In the same way that we can add up alleles for one SNP to create a genotypic score, we can also add up alleles for many SNPs to create a polygenic score, just as we add questionnaire items to create a shyness score.

The results from genome-wide association studies are used to select SNPs and to assign weights to each SNP. For example, in the GWA analysis of weight, the FTO SNP accounts for much more variance than other SNPs, so it should count for much more in a polygenic score for weight.

The following table shows how one individual’s polygenic score is created from ten SNPs. For the first SNP, this individual’s genotype is AT. For this SNP, the T allele happens to be the increasing allele that is positively associated with the trait. So, the individual’s genotypic score for this SNP is 1 because the genotype has only one increasing T allele.

Across the ten SNPs, the individual has a total of nine increasing alleles for the trait out of a possible score of 20. So, this individual would have a polygenic score just below the population average score of 10 for this trait.

This score merely adds the number of increasing alleles, which works reasonably well as a polygenic score.



However, we can increase its precision by weighting the genotypic score for each SNP by how much the SNP correlates with the trait. The correlation between each SNP and the trait is taken from the GWA analysis. If one SNP correlates five times more with the trait than another SNP such as SNP 1 versus SNP 10 it should count for five times as much in the polygenic score.

The weighted genotypic scores in the last column of the table are the product of the genotypic score for each SNP and the correlation with the trait. The sum of these weighted genotypic scores for the ten SNPs is 0.023.

This number isn’t as interpretable as the unweighted genotypic score of 9, which is just the sum of the ‘increasing’ alleles. However, both the unweighted polygenic score of 9 and the weighted score of 0.023 can be expressed simply as a percentile in the population. For this individual, both types would indicate a polygenic score just below average.

How many SNPs should go into a polygenic score? Initially, polygenic scores were created using only the genome-wide significant ‘hits’ from a GWA study. For weight, ninety-seven independent SNPs reached genome-wide significance. Creating a polygenic score from these top ninety-seven SNPs explains 1.2. percent of the variance in weight in independent samples. This is only slightly better than the prediction from the FTO SNP by itself, which explains 0.7 per cent of the variance.

Using only genome-wide significant hits is like demanding that each item in our shyness scale predicts significantly on its own. We don’t do this for other psychological scores because it is unrealistic to expect each item to stand on its own. The goal is to have a composite scale that is as useful as possible.

A better idea is to do what we do when we create other psychological scores: keep adding items as long as they add to the reliability and validity of the composite in independent samples. For polygenic scores, the key criterion is prediction. The new approach to polygenic scores is to keep adding SNPs as long as they add to the predictive power of the polygenic score in independent samples.

This is the strategy that has paid off in the last two years in producing powerful polygenic scores for psychological traits. Some false positives will be included in the polygenic score but that is acceptable as long as the signal increases relative to the noise, in the sense that the polygenic score predicts more variance.  ...

To interpret polygenic scores, it is important to keep in mind that they are always distributed like a bell-shaped curve, that is, a normal distribution. This bell-shaped curve is dictated by the fundamental law of probability, the central limit theorem, which is the basis for all statistics.




The normal distribution is found when many random events contribute to a phenomenon, like flipping a coin and counting the number of times the coin comes up heads. If you flip a coin ten times, you could get no heads or ten heads in a row, but most of the time the total number of heads will be between four and seven. If you do this many times, you will get a perfectly normal bell-shaped distribution, peaking at five, which will be the average number of heads. Flipping coins and counting heads is exactly analogous to counting the numbers of ‘increasing’ alleles from SNPs to construct polygenic scores for many individuals.

I will describe all my polygenic scores in terms of percentiles in the normal distribution. That is, to what extent is my polygenic score above or below the average polygenic score in the comparison sample, the 50th percentile?

It turns out that my polygenic score for height is at the 90th percentile. So, based on my DNA alone, knowing nothing else about me, you could predict that I am tall. And, in fact, I am 6 feet 5 inches. Of course, you can easily see that I am tall if you saw me, but with DNA you could tell that I am tall without even looking at me.

Most importantly, you could have predicted when I was born that I would be tall. Unlike any other predictors, polygenic scores are just as predictive from birth as from any other age because inherited DNA sequence does not change during life. In contrast, height at birth scarcely predicts adult height.

The predictive power of polygenic scores is greater than any other predictors, even the height of the individuals’ parents. Another advantage of polygenic scores over family resemblance is that parental height provides only a family-wide prediction that is the same for any child born to those parents.

In contrast, polygenic scores provide a prediction specific to each individual. In other words, my polygenic scores at birth would have predicted that I would be taller than expected on the basis of the average height of my parents.

Before looking at my other polygenic scores, one other general point needs to be highlighted about predicting individuals. My actual height is at the 99th percentile but my polygenic score is at the 90th percentile. Are polygenic scores sufficiently accurate for prediction?

For example, in TEDS [Twins Early Development Study - 1994 onwards], the polygenic score for height predicts 15 percent of the variance in actual height in these young adults. But 15 per cent is a long way from 100 per cent.

In fact, polygenic scores can never predict 100 per cent of the variance of any trait, because the ceiling for prediction is heritability. For height, heritability is 80 percent, but for psychological traits heritability is 50 percent, which means that polygenic score prediction is always going to be way south of perfect.




The big question is the extent to which polygenic scores will be able to predict all the heritable variance of traits. This gap is called missing heritability, and is described in the Notes section at the end of this book. "

--- [end of text extract] ---

Plomin's scatter plot looks rather messy but it hides the extent of clustering when you consider each decile separately.



As mentioned in the legend above, the vertical lines indicate the 95% confidence intervals. They clearly illustrate the linear trend line. Also, I suspect the limitations on sample size (20,000 pairs of twins in TEDS) for genetic studies.

For the top and bottom deciles, the following chart shows the extent of overlap.


Yet at the extremes the differences are very large. This is a general truth as regards traits whose values are are normally distributed.

Plomin finishes this section by emphasising yet again that due to the current lack of power (too few SNPs identified) and the ceiling of heritability, the PGS prediction is just that - a prediction with an error distribution around it. It is not deterministic.

I would comment that the error bars may be long right now, but as sample sizes get larger and non-additive effects are factored in, they can be made considerably smaller.

This is a small excerpt from an excellent book, by the way, which I reviewed here.

Thursday, November 22, 2018

Every odd number is the difference between two squares

From here via SSC. This is apparently a 'twitter' proof (ie short).



Algebraically (n + 1)2 - n2 = 2n + 1 which is odd.

A slight defence of homo economicus

Amazon link

I've just started this and on page 5 I read:
" Many times the information required to arrive at the theoretical optimum is not available to those who make the decisions, and even if the information is available, the people who make the decisions often do not have an incentive to implement the optimal policy."
If we could only be more like homo economicus, how much better capitalism would work. Cue the application of behavioural genetics in engineering mode.

---

Homo economicus is "... a portrayal of humans as agents who are consistently rational and narrowly self-interested, and who usually pursue their subjectively-defined ends optimally." (Wikipedia).

In fact public choice theory does assume homo economicus, but asserts that incentives often can't or don't align - eg the principal-agent problem, market failure, etc. In capitalist economies consisting of real people, things work even less well.

Who or what should we blame?

Wednesday, November 21, 2018

"Blueprint" - Robert Plomin (a review)

Amazon link

Plomin's book is written in a practical, down to earth style. It's not at all academic. The big ideas (relentlessly hammered home) are these.

  • All traits (physical and psychological) have a substantial genetic component. Heritability is around 50% as a rule of thumb. Higher for traits like height, weight and intelligence.

  • Shared environments such as the family and school do not significantly affect traits. Schools don't make you smarter and parents don't make you nicer.

  • Schools will educate you to your potential aptitude if they're any good (and may socialise you into an elite). Bad parenting can damage a child. But absent active damage, it's the child's inherited genetics which determine performance and personal outcomes.

  • Non-genetic influences on traits are the generally random effects of life events and short lasting (these include test imprecision).

  • Apparent environmental differences (eg parents who read to their children in a house full of books as against ...) have a substantial genetic component (studious parents have studious children). Plomin describes this as nature in nurture.

Strong claims, but based on vast experiments with hundreds of thousands of participants. The results strongly replicate. Problem is, they're quite counterintuitive and few people believe them.

Toby Young writes in Quillette about a recent debate in London:
"On Monday in London’s Emmanuel Centre a debate took place that pitted two Quillette contributors - Robert Plomin and Stuart Ritchie - against two “experts” on child psychology - Susan Pawlby and Ann Pleshette Murphy. The motion was “Parenting doesn’t matter (or not as much as you think)” ...

The ushers asked people to vote for or against the motion on their way in and then again at the end, the idea being that the “winners” would be the side that persuaded the most people to change their minds rather than the side that got the most votes. Which was just as well for Plomin and Ritchie since only 17 percent agreed with them at the beginning of the evening, with 66 percent against and 17 percent saying “Don’t Know.”
Young describes the debate in detail and then describes the results:
"The debate was well-chaired by Xand van Tulleken, a doctor and broadcaster who has an identical twin brother named Chris, and, after he’d taken plenty of questions and done his best to sum up, the audience was asked to vote again.

As expected, a majority still disagreed with the motion, but Plomin and Ritchie had succeeded in persuading some people to change their minds. The number against the motion declined from 66 percent to 51 percent, while those in favor increased from 17 percent to 29 percent, with 20 percent saying “Don’t Know.”

That made Plomin and Ritchie the winners."
Few people truly believe scientific abstractions until the engineering stares them in the face. But that could happen, given the falling cost of whole genome sequencing together with ever larger scale Genome-Wide Association Studies (GWAS) for every trait imaginable.

Soon anyone's genome will be cheaply sequenced and their polygenic scores read off (at any age including infancy, in utero or for IVF selection) to deliver personalised results for a palette of physical, psychological and aptitude life-traits.

Already there are the early signs that the liberal media, the op-ed writers and professional pundits are beginning to warm to the idea. Plomin seems to have avoided the public evisceration he was undoubtedly fearing.

The book is interesting, important and enlightening. One of the books of the year.

---

Eric Turkheimer on his blog is complaining that Plomin has used his ideas without attribution. This may be true but he should get over it: this is popularisation, not academia. If the reception had been super-hostile, Turkheimer might now be experiencing relief rather than angst.

Turkheimer is not alone, by the way. Few of Plomin's peers get namechecked. There could be a few more bruised egos out there.

---

James Thompson of UCL has an excellent and detailed chapter-by-chapter review and Greg Cochran wrote a review at Quillette, "Forget Nature Versus Nurture. Nature Has Won".

Worth checking both out.

Tuesday, November 20, 2018

The Banach–Tarski paradox



According to Wikipedia:
"Given a solid ball in 3‑dimensional space, there exists a decomposition of the ball into a finite number of disjoint subsets, which can then be put back together in a different way to yield two identical copies of the original ball. Indeed, the reassembly process involves only moving the pieces around and rotating them without changing their shape.

However, the pieces themselves are not "solids" in the usual sense, but infinite scatterings of points. The reconstruction can work with as few as five pieces.

A stronger form of the theorem implies that given any two "reasonable" solid objects (such as a small ball and a huge ball), the cut pieces of either one can be reassembled into the other. This is often stated informally as "a pea can be chopped up and reassembled into the Sun" and called the "pea and the Sun paradox".
I remember reading Richard Feynman's reaction to this in his memoir "Surely You're Joking Mr Feynman!",
"Then I [Feynman] got an idea. I challenged them [the mathematicians]: “I bet there isn’t a single theorem that you can tell me – what the assumptions are and what the theorem is in terms I can understand – where I can’t tell you right away whether it’s true or false.”

It often went like this: They would explain to me, “You’ve got an orange, OK? Now you cut the orange into a finite number of pieces, put it back together, and it’s as big as the sun. True or false?”

“No holes.”

“Impossible!

“Ha! Everybody gather around! It’s So-and-so’s theorem of immeasurable measure!”

Just when they think they’ve got me, I remind them, “But you said an orange! You can’t cut the orange peel any thinner than the atoms.”

“But we have the condition of continuity: We can keep on cutting!”

“No, you said an orange, so I assumed that you meant a real orange.”

So I always won. If I guessed it right, great. If I guessed it wrong, there was always something I could find in their simplification that they left out."

There are no infinities in reality.

The LTOV including rent answers critics



The Labour Theory of Value (LTOV) has not had a good press from neoclassical economists.
"However, Ricardo was troubled with some deviations in prices from proportionality with the labor required to produce them. For example, he said "I cannot get over the difficulty of the wine, which is kept in the cellar for three or four years [i.e., while constantly increasing in exchange value], or that of the oak tree, which perhaps originally had not 2 shillings expended on it in the way of labour, and yet comes to be worth £100."   [Wikipedia].
As we shall see, the solution to this mystery is the combination of socially-necessary labour time and rent.

---

Q. What is value theory in Marxism?

A. Marx distinguished three kinds of value. Firstly value itself, the amount of (socially necessary) labour time which went into the production of an object; secondly the exchange value of the produced object when it was exchanged in a market transaction for another object (which could be the universal commodity money); thirdly and qualitatively, the use value which as its name implies was the utility of the object to a person.

Q. Tell me more about value per se

A. Take a wooden box made by a carpenter by hand. Plainly a lazy, incompetent carpenter would take longer to make the box but that extra labour time could not increase its value. A given society knitted together by market transactions would soon arrive at the notion of the average period of time (hours, say) which a box like that would take to produce. That time is the socially necessary labour time. If productivity improved, the value would go down, not up.

Q. And exchange value?

A. An object which is exchanged in a market transaction is called a commodity. A commodity will typically exchange in barter for another object which took something like an equivalent time to produce (otherwise someone is getting a free ride .. and eventually more people will move into producing that commodity and compete). In any kind of competitive and established market a commodity typically exchanges for the money commodity, which makes transactions so much more fluid. The exchange value of a commodity is then its equivalent in money, which in a very simple model would be almost the same concept as its price.

The idea is that in a competitive market of freely working producers the market price will tend to oscillate - due to vagaries of supply and demand in the first instance - around the value.

It's a first, simple model.

Q. And use value?

A. This is not measured in hours or currency units. It's not a quantitative attribute. It instead registers the qualitative fact that anything socially produced has to have utility for someone, otherwise why did the producer bother. The non-trivial idea is that objects can be produced, say for the use value  of household consumption, which have a value (because they took a certain amount of time to do or make) but don't have an exchange value because they were never traded.

Q. So here's a scenario. A tropical island. The guys are doing hunting and fishing and some hard-scrabble cutivating so there's an economy and a currency of sorts. 

And then this big guy monopolises the one set of banana trees and extorts payment for bananas. How does that work in terms of value?

A. Good example. Bananas have a use value, obviously, otherwise everyone would just ignore the big guy. By hypothesis, bananas make themselves. No-one has to labour (by assumption) so the value of a banana is zero. Likewise the exchange value of a banana is also zero - as no labour is incorporated in it.

Q. But the big guy is able to charge. He makes money doesn't he?

A. So this is a case where price and exchange value differ. The big guy is extracting a rent due to his monopolising a scarce resource. We see the same with landowners, who also charge rent to access land which they have perhaps done nothing at all to improve.

If the big guy wasn't there, the bananas would be a 'windfall' and it would be first-come first-served, absent a state structure to ration and allocate (like the land runs in the nineteenth century USA).

Note that value theory can't predict how big the rent (ie the market price) will be. Who knows how much people will value a banana over other traded objects. How much disposable income they will have to allocate to bananas.

Here's the important bit. Some commodities mix aspects of value creation and rent, often when a natural process is involved such as in forestry or wine maturing. The ability of the landowner or wine-make to monopolise the resource while nature does its work allows the extraction of a rent.

This addresses Ricardo's difficulty mentioned at the top of this piece.

Q. Rent vs exchange value - a distinction without a difference?

A. No. Exchange value results from value creation; rents are simply value appropriation.

Suppose we had a guy who guards bananas (which grow themselves), a guy who guards oranges (which grow themselves), a guy who guards a bay with super-abundant fish stocks (which collect and beach themselves). Let's assume they barter. What are the right ratios for exchange?

Marx would say that all three products are not commodities, they have no value. The prices are rents - set by force majeure and need.

Assuming all three parties require all three products then if each is in local abundance the price will be zero. Each is wasting their time trying to monopolise their resource. It's like the air guys. Chill!

If the products are not in abundance only force will decide. The most aggressive guy will set whatever barter ratio he can enforce, then the next most aggressive. However, the underserved guys will eventually die of starvation, making this self-defeating. Hoarding and rationing may postpone the fateful day.

In microeconomic terms, the supply cannot be changed since it is constrained by nature (it makes itself) so the supply curve is vertical. In the absence of substitutes the demand curve is vertical too: to survive you do need a certain level of resources.

If the two vertical lines coincide we have an equilibrium (of sufficiency). If the supply line is to the right  everyone lives and some produce will rot. If it's to the left the guys will become increasingly malnourished and eventually will die.

Take a look at the diagram at the top of the post.

This simple model is also relevant to the owners of entirely robotic factories under total automation, but that's for a future post..

Sunday, November 18, 2018

The cheesecake paradox

This afternoon Clare told me: "Tonight, you'll either have a cheesecake for dessert, or you won't. It'll be a surprise."

OK, she was toying with me. She knows I am torn between desire and calories. I mentally charted my options like this.

Figure 1

But then she had second thoughts. "No," she said. "I do want you to have the cheesecake, but I also want it to be a real surprise. You'll definitely get the cheesecake tonight or tomorrow night. One of the two. But I won't tell you which."

OK, I was disappointed, I have to admit. I hate even the possibility of  delayed gratification. A new diagram took shape in my mind.


Figure 2

I'd either get the cheesecake tonight or it would be the empty plate, in which case at least I'd be getting cheesecake tomorrow.

But wait. It's meant to be a surprise. The surprise is all in the branching possibilities. If I didn't get cheesecake tonight, then I'd be in the situation below (yellow oval).

Figure 3

No surprise there! There's no branching in the yellow oval. So there's no surprise tomorrow. And that must mean I'm getting the cheesecake tonight! Excellent!


Figure 4

Oh, wait again. No branching here. So my very certainty has removed the element of surprise. So it looks like Clare's proposal to give me a surprise cheesecake has collapsed. There's no way she can do it.

No cheesecake at all. Cruel.

Figure 5

But now I'm sure I'll get nothing, I can't rule out the possibility that nevertheless she may give me a cheesecake tonight after all.

Figure 6 = figure 2

And I will be very surprised!

---

I wrote about this paradox back in 2009. I noted that despite its simplicity, according to Wikipedia, no-one has a good solution. It looks to me that the self-reference feedback is creating alternate, flip-flopping states of certainty and uncertainty.

Another example of how concepts from AI such as game trees, the perspective of an agent with cognitive states, and Kripke models of doxastic logic can illuminate problems which seem intractable from a purely logicist or philosophical standpoint.

Friday, November 16, 2018

Mathematical determinism: 2 + 2 = 4

This mini-rant on reading how Robert Plomin's book has been traduced by the usual suspects: "Genetic determinism rides again".

---
  • "So 2 + 2 = 10 (base four), 2 + 2 = 1 (mod 3). So much for determinism."

  • "Does '2 + 2' look anything like '4' to you?" (Rolls eyes).

  • "Western imperialist bias: ٢ + ۲ = ٤."

  • "If you had two apples and two oranges that's not four of anything."

  • "So two monogamous couples make a foursome? In your dreams, chauvinist!"

  • "Like white middle-aged males reduce everything to abstractions!"

  • "The voice of privilege. This is so not true for disadvantaged minorities."

  • "It depends on what '2', '+', '=' and '4' are taken to mean. They're just conventions reflecting the patriarchal power structure."

  • "An arrogant assertion presented without a shred of proof."

  • "Life is more complex and interesting than sterile theories. Get over it."

  • "Nazis/Communists/Zionists/Russians believe that."

  • "You're just saying that to make me feel uncomfortable. You're so not in touch with your own feelings."

  • "We shouldn't discriminate against people who have a problem with what you're saying."

etc etc.

---

Say something true at the margins of the Overton window and be pecked to death by spurious attacks until everyone has forgotten the original proposition but is left believing it has been convincingly refuted.

---

It's an interesting exercise, a homework assignment, to identify the fallacies in each of the 'rebuttals' above. The required concepts can be quite sophisticated: initial algebra; the formalisation of syntax and semantics in mathematical logic - wffs, valuation functions and models; the formal notions of proof and interpretation in mathematics and science.

Add a dash of sophistry and emotionalism to complete.

Wednesday, November 14, 2018

"Blueprint: How DNA Makes Us Who We Are" - Robert Plomin

Amazon link

Robert Plomin is one of the good guys, and this book is apparently a summing up, for a general audience, of his life work. It has just arrived and is on the stack.

Meanwhile Dr James Thompson of UCL has an excellent and detailed chapter-by-chapter review, from which this short excerpt.
"Chapter 11 is about the development of genome-wide association studies. Chapter 12 is a very substantial one about genetic prediction. Chapter 12 is Plomin’s real coming out: he reveals his polygenic scores for all to see. Naturally, this is a teaching opportunity, explaining the insights and the limitations of such measures. Figs 6 and 7 are worth showing again and again, if only to explain polygenic scores and their overlaps, and the fact that they provide probabilistic estimates, not certainties."
Greg Cochran has written a review at Quillette, "Forget Nature Versus Nurture. Nature Has Won".

And Toby Young, also at Quillette: "Is Sociogenomics Racist?".  No, he argues.

I'm looking forward to reading and learning. Update: here's my review.

Monday, November 12, 2018

A feminist on sex machines: Kate Devlin

Amazon link

Dr Kate Devlin's Wikipedia entry.

---

The concept of a 'machine for sex’ is a wondrous, powerful one. As the author notes, its arc aligns with all of human history. Devlin describes ‘sex toys’ from prehistory and antiquity right through to the present day.

The term ‘sex robot’ itself is rather reified. Better is the idea of an artefact with which humans may engage sexually. Such an artefact might be genetically engineered, machine fabricated or some hybrid in between. Sex devices interestingly inhabit the intersection of science and technology, genetics and evolutionary biology, and the myriad social sciences.

If a sex robot is defined as a thing, as human property, then they existed once, currently don't really exist and may exist again in the middling future. In antiquity slaves were property and presumed to have no agency. It was routine and uncontroversial for elite slave-owning males to buy and use nubile male and female slaves for sex. As Kyle Harper recounts in his “Slavery in the Late Roman World, AD 275–425”, consequences included masters becoming (stupidly) besotted with their slave sex-objects, wifely jealousy .. and unintended offspring who would join the next generation of slaves.

Antique ‘sex robots’ were, from a functional point of view, poorly implemented. Their inner drives and motivations were not aligned with their 'function’ which made ownership fraught and only manageable through sustained terror.

Devlin’s first degree was in archaeology so she will know all this. It would have been good to read about the social, ethical and even practical implications of the widespread availability of high-functioning ‘sex robots’ in antiquity. Devlin's discussion is however (pp.116-117) superficial, flagging only the usual oppressively gendered roles found in all premodern societies stabilised by male violence. She also notes a pre-Christian sexual disinhibition of which she approves.

“Turned On” is not a book of science with some feminist advocacy. It is instead a feminist tract anchored around the topic of sex-with-artefacts (p.213). What’s this for example - a mocking rebuke to the transgression of equal outcomes?
“Why do we experience things? How do the mechanisms of our bodies and brains give rise to conscious sensations? Where does that consciousness come from? You don’t have to have an answer in the Great Zombie Debate. The philosophers can’t agree on it either, Which is why you have a bunch of very clever middle-aged white men amusingly inventing words like‘ ‘zoombie’ and ‘zimboe’ to put forward their own variations on the theory. “ (p. 102)
My emphasis. It's certainly of the moment, but it jars.

Devlin tells us she is a feminist, an ideology developed to further the interests of liberal professional women like Dr Devlin. Like all ideologies, it’s protean and eclectic, cherry picking arguments which support the cause. A characteristic of ideologies is that they narrate a world their proponents wish they were in but which rarely coincides with actuality or realistic possible outcomes.

Consider objectification, something she takes strong issue with. The discussion is phenomenological: an example might be a builder wolf-whistling an attractive woman in the street; or a bunch of women hooting and laughing at a male stripper at a ‘hen party’.

In the broad-brush triune brain model, primary drives such as lust are associated with the (animalistic) brainstem formation; emotional attachment with the mammalian limbic system; and rational interaction with hominid cortical systems. It’s not much of a surprise that humans can exhibit sexual behaviour dominated by any of these loci.

Objectification (the mode of lust-dominance) would then be a brain-stem determined form of behaviour. In more refined circles we expect cortical inhibition/mediation of primary drives to deliver more measured, prosocial behaviour factoring in social context. Our biology is complicated in social settings.

There isn’t any such framing in Devlin's book. Just normative presumption that objectification is out there (for some reason), that’s it’s wrong and that current sex doll designs play up to it. This is to sell the reader short, replacing analysis with moralising.

Naturally we don't wish to succumb to the naturalistic fallacy; humans with their small group evolutionary history are imperfectly adapted to large scale societies. There are plenty of natural urges we need to regulate and indeed legislate against. There are few easy answers here as Devlin would be the first to argue, in contexts such as the legality and ethics of child sex dolls.

Devlin adopts the fashionable feminist view that phenomenal gender differences are purely social constructs. This despite the enormous weight of hard evidence (evolutionary, neuroanatomic, genomic, physical, psychometric) for well-defined and reproducible biological differences between the sexes. Differences which are hardly obscure, but recognition of which might undermine the claims of her interest group. Public choice theory assumes its usual relevance here.

Devlin interviews the CEO of RealDoll, Matt McMullen, and gives him a hard time about the overwhelming preponderance of hyper-sexualised female dolls and robots. 'Where are the less-sexualised dolls, the male dolls?' she wants to know. Actually this is a point she takes up with all the doll manufacturers she meets. The replies she gets are defensive, framed in terms of male-female differences in sexuality leading to skews in demand. Devlin is having none of it, blaming biased marketing and uncritical social conditioning (pp. 153-154). Time for a quick review of microeconomics (supply-demand equilibria?) then as I reflect on the democracy of markets in probing the world as it actually is.

Then we read this in a meandering discussion of rape fantasies and the claimed lack of any genetic influence:
“If anything, rape would theoretically reduce the reproductive success of our ancestors as it takes away selective genetic choice. “ (p. 233).
Reality is more nuanced. Rape is historically (and currently) commonplace in intergroup conflicts. It's plainly adaptive for males in the absence of draconian ingroup penalties. I suggest a quick read of Dawkins' "The Selfish Gene".

Gendered social roles have historically been oppressive to women. There are biological reasons (eg male physical strength, aggression and paternity-uncertainty) which interlink with social reasons (eg within-family inheritance) which are specific to different kinds of society and which need to be explicitly teased out. If capitalism appears to be intrinsically gender-blind in its desire to free everyone up to maximally work, then the deleterious effects on human self-reproduction also need some analysis.

A perennial science-fictional trope referenced by Tom Whipple, The Times science editor in his review, is that of the perfected sex robot as a kind of sterile mosquito (which has already resulted in some local extinctions of this malaria disease vector).

This is not a problem we'll face anytime soon but is there something to it? Devlin is unworried while Whipple remains concerned. Plainly it’s hard to assess an unknown artefact but with universal and easy access to contraception in the west, perhaps we already have a natural experiment. Check those Total Fertility Ratios.

We are already selecting for women who positively want children (rather than just sex); ubiquitous effective and sterile sex robots would select for men with a similar drive. Let's hope there's that much variability in the gene pool.

In summary Devlin's book is an easy and amusing read: somewhat superficial; an interesting tour of the sex doll/sex toy landscape which will be unfamiliar to many readers; intriguing confessional snippets from the author's private life.

The book is not particularly scandalous or salacious, it's not very conceptual or analytic and its opinions are conventional liberal left. Put aside a slightly sprawling and uneven structure and it reads like an extended New Scientist article.

Incidentally, if Dr Devlin or her co-workers were ever to read this review, they would not be won over. It is in the nature of ideologies to be believed by their adherents and they provide powerful mechanisms of contextualisation and framing to theorise their opponents. This review would be framed as a defence of the patriarchal power structure: discredited gender essentialism.

We are truly the victims of our own axioms.

---

Further reading: "So Beautiful".

Sunday, November 11, 2018

A second review of "The Master Algorithm" - Pedro Domingos

Amazon link

This is a guest review from Dr Roy Simpson.

---

Review of The Master Algorithm (by Pedro Domingos) 

By Dr. R. Simpson

This book provides both a history and a visionary project in machine learning by a leading professor in the field. The author writes fluently and well providing an informative book, which can be worth re-reading if one is interested in the details of machine learning techniques, as well as in (re-)evaluating his ideas about the future of the subject.

I was brought to this book after reading two other recent books on the subject of Algorithms: Weapons of Math Destruction, by Cathy O'Neil and the recent Hello World by Hannah Fry. Both of these books discuss the ongoing issues associated with the application of Algorithms, Big Data and Machine Learning to contemporary society – the latter book being newer and most relevant for UK audiences. These books recommended The Master Algorithm (2015) as a book for a deeper understanding of machine learning.

Much of this book is eminently quotable, and its text provides good introductions, for example here is from the first page of the Prologue:

You may not know it, but machine learning is all around you. When you type a query into a search engine, it's how the engine figures out which results to show you (and which ads as well). When you read your email, you don't see most of the spam, because machine learning filtered it out. Go to Amazon.com to buy a book or Netflix to watch a video, and a machine-learning system helpfully recommends some you might like...

Traditionally, the only way to get a computer to do something – from adding two numbers to flying an airplane - was to write down and algorithm explaining how, in painstaking detail. But machinelearning algorithms, also known as learners, are different: they figure it out on their own, by making inferences from data. And the more data they have, the better they get. Now we don't have to program computers; they program themselves.

It's not just in cyberspace, either: your whole day, from the moment you wake up to the moment you fall asleep, is suffused with machine learning.

The introductory chapters continue with more aspects of machine learning making some interesting points. For example the above passage indicates a “culture shift” within the programming world. In a later section it is noted that Microsoft has some difficulties with the new world, because its programmers are just that and its main products are produced in the traditional way, whereas Google is more of a machine learning organisation, with its main products produced the machine learning way.

He introduces the metaphor that whereas traditional programming is more like a manual (at best industrial) process, machine learning is more like farming – prepare the ground, then sit back and watch the systems (i.e. commercial products) grow themselves.

The author is aware that this field is not “General AI” - a topic often discussed in this blog to which I shall return at the end of this review. However we read that within this subfield of AI the traditional opponents are “Knowledge Engineers”. Knowledge Engineers hold (or at least once held) the view that systems which contain “knowledge” need to have that knowledge typed into them, and more generally be programmed in the traditional way.

The classic example was Cyc - a massive common sense knowledge based system over which programmers have been typing in “common sense rules” for several decades now.

The main counter to this view presented in this book is the matter of scale: knowledge engineers could work with thousands of rules; whereas machine learning will generate many millions of rules. (The argument against this knowledge engineering viewpoint held by the Agent Theory/AGI community has been different, and concerns its narrowness of scope. Agent Theorists might have used a “narrowness of scope” argument against early machine learning too, but the scalability of these techniques has won the commercial argument  - at least for now - though not necessarily the conceptual argument, as discussed later in this review.)

The primary content of the book is a review and overall analysis, from a modern machine learning perspective, of the five main strands of machine learning during the history of AI: Symbolism, Connectionism, Evolutionary Programming, Bayesian Networks, and Analogical Reasoning.

His purpose in all of this is to identify the principles involved and describe the essence of the technique and identify the best algorithm that the given technique has provided to the machine learning community. From there he moves to the main objective of the book: the development of a “Master (Learning) Algorithm” which incorporates all the best of the previous techniques and which is the subject of his own research group. This Master Algorithm would then be able to optimally learn anything .. .

A brief summary of each (with quotes from the book):

Symbolism

The symbolist's core belief is that all intelligence can be reduced to manipulating symbols. Whereas deductive reasoning is about going from axioms to conclusions, the learning aspect requires inductive reasoning: going from conclusions to axioms. Thus Hume's problem is discussed and the master algorithm here becomes “inverse deduction”.

Inverse deduction is like a super-scientist systematically looking at the evidence, considering possible inductions, collating the strongest, and using those along with other evidence to construct yet further hypotheses – all at the speed of computers. Yet this inverse deduction has limitations and issues, so we move to Connectionism.

Connectionism

How does your brain learn? Hebb's rule (from 1949) has become the cornerstone of connectionism, and is about neuron firing: “neurons that fire together wire together”. This history in AI begins with the Perceptron (from the 1950s). This was an electromechanical learning machine, based on a model of neurons, which had some basic classification skills (e.g is this a picture of a door or not?).

In the late 1960s it was proven that its classification skills were too limited to be of much generality and this approach to AI learning suffered a near total set-back (to the delight of the knowledge engineers of the time, due to funding competition).

It was not until the 1980s that physics inspired alternatives were introduced, such as the Boltzmann machine, which introduced probabilities into the field, and brought in techniques from statistical physics (ie thermodynamics) – so for a period the notion of “temperature” was important for a neural network learning system!

Although the introduction of probabilities has lasted, it is not clear what has happened to "temperature" in machine learning. It later transpired also that none of this physics theory was necessary to overcome the original Perceptron limitations. Nevertheless more mathematical techniques emerge from this era such as calculations in hyperspace, with associated weighted functions, convergence metrics, etc.

The main algorithm eventually identified from Connectionism is backpropagation, whose refinements are at the core of today's learning systems. However, the author raises an intriguing question: Is everything we “know” actually learned by our neurons? Has not evolution played a part too?

Evolutionary Learning

As an introduction to this chapter the author tells the following fantasy story:

Robotic Park is a massive robot factory surrounded by ten thousand square miles of jungle, urban and otherwise. Ringing that jungle is the tallest, thickest wall ever built, bristling with sentry posts, searchlights, and gun turrets. The wall has two purposes: to keep trespassers out and the park's inhabitants – millions of robots battling for survival and control of the factory – within.

The winning robots get to spawn, their reproduction accomplished by programming the banks of 3D printers inside. Step-by-step, the robots become smarter, faster – and deadlier. Robotic Park is run by the US Army, and its purpose is to evolve the ultimate soldier.

Needless to say this story brings out a large number of issues and concerns. The author seems to be sanguine about the dangers here, although modern “AI Ethics” movements (which included the late Stephen Hawking and Elon Musk) are concerned about this type of development. Towards the end of the book, the author actually suggests that working on “AI Ethics” could itself be a growth industry for displaced humans in the new era – robots and AI will have a lot of (human) ethics to learn!

The author is using this story here to dramatically introduce the AI Learning techniques inspired by Darwinian evolution: genetic algorithms and genetic programming. These techniques are overviewed, and some issues identified. However this reviewer is intrigued by the wider point that is being made here, and has an alternative way of expressing a related idea. The overall point that the author is making can be summarised by this formula:

   Learning == Agent Learning + Environment Learning + (Environment → Agent transfer)

In other words, when we ask “how does that small brain learn all this stuff?” - the answer is that the small brain has not had to do all the learning implicit in its actions. This viewpoint is implied also by the Chomsky view of language acquisition (Chomsky has been another critic of the traditional approach to machine learning – the author hopes that his wider “Master Algorithm” approach meets Chomsky's concerns.)

There are again several limitations to any master algorithm provided by what we could also call the Darwinian algorithm. Chief amongst them is the fact that the Darwinian algorithm (as currently understood) tends to find “suboptimal” solutions, not optimal solutions – so we move on to Bayesian theory.

Bayesian Networks

The path to optimal learning begins with a formula that many people have heard of: Bayes Theorem. But here we'll see it in a whole new light and realize that it's vastly more powerful than you'd guess from its everyday uses.

At heart, Bayes' theorem is just a simple rule for updating your degree of belief in a hypothesis when you receive new evidence: if the evidence is consistent with the hypothesis, the probability of the hypothesis goes up; if not, it goes down.

Bayes Theorem uses conditional probabilities. (P(A|B) is the probability that A happens given that B has happened/is assumed) and there is an associated inference system. Similar to deductive logic (e.g. the Prolog-based resolution systems sometimes discussed in this blog) the inference system allows the deduction of probabilistic conclusions and the management of probabilistic assumptions. This is all wrapped into a network structure amongst assumptions, and apparently it has been discovered that there is an isomorphism between the probabilities and weights often used in Artificial Neural Net models and the probabilities used in Bayesian Networks.

The success of this Bayesian approach arises directly from the quantity of data now available, making the probabilities and conditional probabilities very accurately determinable from millions upon millions of data points (pre-Internet such probabilities would have been input by researchers by hand as guesses, resulting in unconvincing performance and results from such probabilistic systems).

This has motivated much of the idea behind a complete Master Algorithm for learning. This has left open two remaining unification tasks, of which the hardest has been the unification of logic and probability; the other has been to incorporate what to do when there is essentially no data, only analogies to existing data.

Analogy Reasoning

Methane and Methanol have very similar chemical structure; but they are not identical and have some big differences since one is a (room temperature) gas, the other a liquid. So if your system knew much about one, what can it deduce about the other? Much of science and business-services-work operates by analogy: no two customers are exactly the same. We manage to cope, so what are the principles involved?

Various classification ideas have been developed in this strand of AI, and a very powerful technique called a Support Vector Machine was apparently the most powerful AI technique around the turn of the century. Only in recent years has it been superseded by Artificial Neural Nets for the top learning slot (although Artificial Neural Nets don't do analogical reasoning as such).

The Master Algorithm - Alchemy

So putting this all together has been the research project of Prof. Domingos, and he has developed a mathematical framework called Markov Logic Networks and a corresponding (open source) software system called Alchemy. The final chapters discuss the general steps required to produce a Master Algorithm from all the above ingredients. He does not describe the current form of Alchemy (or MLN) as the Master Algorithm, but uses it to suggest that this goal is feasible.

One feature he seems to claim Alchemy is lacking is the ability to explain itself fully, also there may still be optimisation issues. There are also some discussions about the social aspects of all this, suggesting that users should form “trusted data unions” to hold the value of their data (a bit like banks holding their money), rather than the current practice of just handing everything over to the big corporations who then benefit from any commercial consequences (and as UK residents are aware, don't pay appropriate tax either).

Analysis: Agent Theory and Computational Mathematics

I shall end this review with a few remarks about this idea of a “Master Learning Algorithm” from the perspective of Agent theory (which is often discussed in this blog); and some ideas about Computational Mathematics (my own interest).

Agent theory takes a more holistic, biologically realistic and physically realist view of AI than one finds in stand-alone techniques like machine learning. Although these machine learning techniques have become (and will remain for another decade at least) at the centre of a business and social revolution, in aiming for such a high goal as a Master Algorithm which can learn everything – and do so optimally - one has to ask if boundaries will eventually be reached due to insufficient attention to agent theoretic issues.

For example, how and in what way, is a biological cell a learning system? Likewise the mysteries of neurological biochemistry are not resolved in an agreed way. Even aspects of the Darwinian algorithm are not fully understood. So it is quite possible that nature has more to teach us about learning.

On the subject of Computational Mathematics and AI much can also be said. A famous critic of the entire AI project (at least as seen as a branch of Computer Science) has been Professor Penrose with his books (from 1989 and 1994). Since that time more results have appeared in the foundations of logic and mathematics which can be viewed as clarifying Penrose's position.

For example using the results of an ongoing project known as ReverseMathematics, this reviewer suspects that much of Applied Engineering Mathematics is not actually computable in the Turing sense. However it nearly is, so many effects are “subtle” - although the inability to predict weather systems accurately beyond (say) five days; turbulence theory; and mathematical puzzles related to physics models suggest these “weakly non-computable” effects may yet be important.

The author himself recognises that an outstanding and unsolved mathematical problem known as the NP-completeness problem is significant to the subject of AI. I shall leave the last words to the author:

The purpose of AI systems is to solve NP-complete problems, which may take exponential time, but the solutions can always be checked efficiently. We should therefore welcome with open arms computers that are vastly more powerful than our brains, safe in the knowledge that our job is exponentially easier than theirs.

---

My own review was posted back in June 2016 and can be found here.