Alignment

“In artificial intelligence (AI) and philosophy, the AI control problem is the issue of how to build a superintelligent agent that will aid its creators, and avoid inadvertently building a superintelligence that will harm its creators. Its study is motivated by the notion that humanity will have to solve the control problem before any superintelligence is created, as a poorly designed superintelligence might rationally decide to seize control over its environment and refuse to permit its creators to modify it after launch” (Wikipedia, retrieved June 2021).

In this realm, understanding how to reward machine learning behaviour so as to develop a “policy” that dictates how the “intelligent agents” do what we want them to, has been supplanted by looking instead at structuring the environments in which these agents will operate. In his book “The Alignment Problem: Machine Learning and Human Values” (2020), author Brian Christian explains why, using the example of ourselves in nature:

“A programmed heuristic like, ‘Always eat as much sugar and fat as you can’ is optimal as long as there isn’t all that much sugar or fat in your environment and you aren’t especially good at getting it. Once that dynamic changes, a reward function that served you and your ancestors for tens of thousands of years suddenly leads you off the rails” (2020, p. 173).

Clues from evolution and child development are now useful to reward designers of robots and artificial intelligence. Beyond specific policies, Christian says “values” must be instilled in these agents using notions of parenting and pedagogy, and in a manner where not only will our actions be understandable to our creations, but so that they act in ways that are transparent to us. He cautions against relinquishing too much control, not to the agents and machines, but to the training models we use for these sorts of purposes, citing Hanna Arendt as to how easily evil can emerge from an ill-conceived but otherwise innocuous template, as the models themselves “might become true” (2020, p. 326).

Given their complex nature, should we wonder whether our intelligent machines might develop some equivalent of emotion? In an essay titled “In The Chinese Room, Do Computers Think?”, science author George Johnson suggests such anomalous behaviour could take the form of “qualities and quirks that arose through emergence, through the combination of millions of different processes. Emotions like these might seem as foreign to us as ours would to a machine. We might not even have a name for them” (1987, p. 169).

How might such artificial emotions arise and what might they be like? As science fiction author Philip K. Dick wonders, will our Androids Dream of Electric Sheep? Do such speculations point to how our very own emotions and thoughts arise, and the factors in our bodies and environments which contribute to their arising?


Christian, B. (2020) The Alignment Problem: Machine Learning and Human Values. New York, United States: Penguin Random House.

Dick, P. K. (1968) Do Androids Dream of Electric Sheep? New York, United States: Penguin Random House – Doubleday.

Johnson, G. (1987), In The Chinese Room, Do Computers Think? in Minton, A. J. & Shipka, T. A. (Eds.), Philosophy: Paradox and Discovery Third Edition (1990), (pp. 156 – 170). New York, United States: McGraw-Hill.

Value

Despite backwards and otherwise misinformed views on eugenics and corporal punishment, nineteenth century author and Congregationalist pastor Newell Dwight Hillis acknowledged the very human fears of the Industrial Revolution in his 1896 book “A Man’s Value to Society“:

“Recently the test exhibition of a machine was successful, and those present gave the inventor heartiest congratulations. But one man was present whose face was drawn with pain, and whose eyes were wet with tears. Explaining his emotion to a questioner he said, “One hour ago I entered this room a skilled workman; this machine sends me out the door a common laborer. For years I have been earning five dollars a day as an expert machinist. By economy I hoped to educate my children to a higher sphere, but now my every hope is ruined.” (1896, p. 64).

As we enter the age of Artificial intelligence (AI), do the skilled nineteenth-century workman’s fearful words find resonance once again? Or, such as back then, are there other factors not presently being considered?

Economist and author Daniel Susskind states in his book “A World Without Work” (2020), that the rapid rate at which AI and automation supplant occupations over the coming years could eventually outstrip the amount of work left available for people to do, potentially leaving many unemployed.

Considering a Universal Basic Income (UBI), Susskind argues that such an approach could still leave people in an existential vacuum and exposed to other risks. One proposed solution is a Conditional Basic Income (CBI), which could provide a guaranteed level of economic security and see people compensated based upon things they enjoy doing, or that are nonetheless essential. These might encompass creative pursuits, education, recreation, providing care-giving services to family members, neighbours, or contributing in some way to community and society.

Another approach Susskind explores is a sharing of state capital which could see people each having their own stock of it, as traditional capital, like an endowment (2020, p. 189).

What other innovative approaches could similarly aim to help narrow economic divisions in society and how might they be realized? Can “Big Tech” play a role? Based on current evidence, Susskind suggests perhaps not, as “software engineers, after all, are not hired for the clarity and sophistication of their ethical reasoning” (2020, p. 210). Further along this idea of trust is the question of should we not also be wary of leaving it up to the “Big State” to look after, given the example of China’s new surveillance-driven “social credit system” where citizens are scored and ranked based on everyday conduct (2020, p. 211). Moreover, does looking at potentially new social landscapes from a purely economic angle not limit our perspective overall? As Susskind writes,

“We have tended to turn to economists, the engineers of contemporary life, to tell us how to relentlessly grow the pie. In a world with less work we will need to revisit the fundamental ends once again. The problem is not simply how to live, but how to live well. We will be forced to consider what it really means to live a meaningful life” (2020, p. 236).


Dwight Hillis, N. (1896) A Man’s Value to Society: Studies in Self-Culture and Character [HTML document]. New York, United States: Fleming H. Revell Company

Susskind, D. (2020) A World Without Work: Technology, Automation, and How We Should Respond. New York, United States: Metropolitan Books