Welcome to the Nexus of Ethics, Psychology, Morality, Philosophy and Health Care

Welcome to the nexus of ethics, psychology, morality, technology, health care, and philosophy

Wednesday, October 31, 2018

We’re Worrying About the Wrong Kind of AI

Mark Buchanan
Bloomberg.com
Originally posted June 11, 2018

No computer has yet shown features of true human-level artificial intelligence much less conscious awareness. Some experts think we won't see it for a long time to come. And yet academics, ethicists, developers and policy-makers are already thinking a lot about the day when computers become conscious; not to mention worries about more primitive AI being used in defense projects.

Now consider that biologists have been learning to grow functioning “mini brains” or “brain organoids” from real human cells, and progress has been so fast that researchers are actually worrying about what to do if a piece of tissue in a lab dish suddenly shows signs of having conscious states or reasoning abilities. While we are busy focusing on computer intelligence, AI may arrive in living form first, and bring with it a host of unprecedented ethical challenges.

In the 1930s, the British mathematician Alan Turing famously set out the mathematical foundations for digital computing. It's less well known that Turing later pioneered the mathematical theory of morphogenesis, or how organisms develop from single cells into complex multicellular beings through a sequence of controlled transformations making increasingly intricate structures. Morphogenesis is also a computation, only with a genetic program controlling not just 0s and 1s, but complex chemistry, physics and cellular geometry.

Following Turing's thinking, biologists have learned to control the computation of biological development so accurately that lab growth of artificial organs, even brains, is no longer science fiction.

The information is here.

Learning Others’ Political Views Reduces the Ability to Assess and Use Their Expertise in Nonpolitical Domains

Marks, Joseph and Copland, Eloise and Loh, Eleanor and Sunstein, Cass R. and Sharot, Tali.
Harvard Public Law Working Paper No. 18-22. (April 13, 2018).

Abstract

On political questions, many people are especially likely to consult and learn from those whose political views are similar to their own, thus creating a risk of echo chambers or information cocoons. Here, we test whether the tendency to prefer knowledge from the politically like-minded generalizes to domains that have nothing to do with politics, even when evidence indicates that person is less skilled in that domain than someone with dissimilar political views. Participants had multiple opportunities to learn about others’ (1) political opinions and (2) ability to categorize geometric shapes. They then decided to whom to turn for advice when solving an incentivized shape categorization task. We find that participants falsely concluded that politically like-minded others were better at categorizing shapes and thus chose to hear from them. Participants were also more influenced by politically like-minded others, even when they had good reason not to be. The results demonstrate that knowing about others’ political views interferes with the ability to learn about their competency in unrelated tasks, leading to suboptimal information-seeking decisions and errors in judgement. Our findings have implications for political polarization and social learning in the midst of political divisions.

You can download the paper here.

Probably a good resource to contemplate before discussing politics in psychotherapy.

Tuesday, October 30, 2018

How Trump’s Hateful Speech Raises the Risks of Violence

Cass Sunstein
Bloomberg.com
Originally posted October 28, 2018

Here is an excerpt:

Is President Donald Trump responsible, in some sense, for the mailing of bombs to Hillary Clinton and other Democratic leaders? Is he responsible, in some sense, for the slaughter at the Pittsburgh synagogue?

If we are speaking in terms of causation, the most reasonable answer to both questions, and the safest, is: We don’t really know. More specifically, we don’t know whether these particular crimes would have occurred in the absence of Trump’s hateful and vicious rhetoric (including his enthusiasm for the despicable cry, “Lock her up!”).

But it’s also safe, and plenty reasonable, to insist that across the American population, hateful and vicious rhetoric from the president of the United States is bound to increase risks of violence. Because of that rhetoric, the likelihood of this kind of violence is greater than it would otherwise be. The president is responsible for elevating the risk that people will try to kill Democrats and others seen by some of his followers as “enemies of the people” (including journalists and Jews).

To see why, we should investigate one of the most striking findings in modern social psychology that has been replicated on dozens of occasions. It goes by the name of “group polarization.”

The basic idea is that when people are listening and talking to one another, they tend to end up in a more extreme position in the same direction of the views with which they began. Groups of like-minded people can become radicalized.

The info is here.

West Virginia Poll examines moral and social issues

Brad McElhinny
wvmetronews.com
Originally posted September 30, 2018

Here is an excerpt:

Role of God in morality

There was a 50-50 split in a question asking respondents to select the statement that best reflects their view of the role of God in morality.

Half responded, “It is not necessary to believe in God in order to be moral and have good values.”

The other half of respondents chose the option “It is necessary to believe in God in order to be moral and have good values.”

“The two big, significant differences are younger people and self-identified conservatives who have opposite points of view on this question,” said professional pollster Rex Repass, the author of the West Virginia Poll.

Of younger people — those between ages 18 and 34 — 60 percent said it’s not necessary to believe in God to have good moral and ethical values.

That compared to 35 percent of those ages 55-64 who answered with that statement.

“So generally, if you’re under 35, you’re more likely to say it’s not necessary to say have a higher being in your life to have good values,” Repass said.

“If you’re older that percentage increases. You’re more likely to believe you have to have God in your life to be moral and have good values.”

Of respondents who labeled themselves as conservative, 73 percent said it is necessary to believe in God to have moral values.

The info is here.

Monday, October 29, 2018

We hold people with power to account. Why not algorithms?

Hannah Fry
The Guardian
Originally published September 17, 2018

Here is an excerpt:

But already in our hospitals, our schools, our shops, our courtrooms and our police stations, artificial intelligence is silently working behind the scenes, feeding on our data and making decisions on our behalf. Sure, this technology has the capacity for enormous social good – it can help us diagnose breast cancer, catch serial killers, avoid plane crashes and, as the health secretary, Matt Hancock, has proposed, potentially save lives using NHS data and genomics. Unless we know when to trust our own instincts over the output of a piece of software, however, it also brings the potential for disruption, injustice and unfairness.

If we permit flawed machines to make life-changing decisions on our behalf – by allowing them to pinpoint a murder suspect, to diagnose a condition or take over the wheel of a car – we have to think carefully about what happens when things go wrong.

Back in 2012, a group of 16 Idaho residents with disabilities received some unexpected bad news. The Department of Health and Welfare had just invested in a “budget tool” – a swish piece of software, built by a private company, that automatically calculated their entitlement to state support. It had declared that their care budgets should be slashed by several thousand dollars each, a decision that would put them at serious risk of being institutionalised.

The problem was that the budget tool’s logic didn’t seem to make much sense. While this particular group of people had deep cuts to their allowance, others in a similar position actually had their benefits increased by the machine. As far as anyone could tell from the outside, the computer was essentially plucking numbers out of thin air.

The info is here.

The dismantling of informed consent is a disaster

David Penner
KevinMD.com
Originally posted September 26, 2018

Informed consent is the cornerstone of medical ethics. And every physician must defend this sacred principle from every form of evil that would seek to dismantle, degrade and debase it. If informed consent is the sun, then privacy, confidentiality, dignity, and trust are planets that go around it. For without informed consent, the descent of health care into amorality is inevitable, and the doctor-patient relationship is doomed to ruination, oblivion, and despair. It is also important to acknowledge the fact that a lack of informed consent has become endemic to our health care system.

This betrayal of patient trust is inextricably linked to three violations: a rape of the body, a rape of the mind and a rape of the soul. The rape of the mind is anchored in a willful nondisclosure of common long-term side effects associated with powerful drugs, such as opioids and certain types of chemotherapy. When a patient starts a chemotherapy regimen, they are typically briefed by a nurse, who proceeds to educate them regarding common short-term side effects such as mouth sores, constipation, and nausea, while failing to mention any of the typical long-term side effects, such as cognitive difficulties and early menopause. It is the long-term side effects that underscore the tragedy of having to resort to chemotherapy, as they can have a devastating impact on a patient’s quality of life, even long after remission has been attained.

The info is here.

Sunday, October 28, 2018

Moral enhancement and the good life

Hazem Zohny
Med Health Care and Philos (2018).
https://doi.org/10.1007/s11019-018-9868-4

Abstract

One approach to defining enhancement is in the form of bodily or mental changes that tend to improve a person’s well-being. Such a “welfarist account”, however, seems to conflict with moral enhancement: consider an intervention that improves someone’s moral motives but which ultimately diminishes their well-being. According to the welfarist account, this would not be an instance of enhancement—in fact, as I argue, it would count as a disability. This seems to pose a serious limitation for the account. Here, I elaborate on this limitation and argue that, despite it, there is a crucial role for such a welfarist account to play in our practical deliberations about moral enhancement. I do this by exploring four scenarios where a person’s motives are improved at the cost of their well-being. A framework emerges from these scenarios which can clarify disagreements about moral enhancement and help sharpen arguments for and against it.

The article is here.

Saturday, October 27, 2018

Obtaining consensus in psychotherapy: What holds us back?

Goldfried, M.R.
American Psychologist
2018

Abstract

Although the field of psychotherapy has been in existence for well over a century, it nonetheless continues to be preparadigmatic, lacking a consensus or scientific core. Instead, it is characterized by a large and increasing number of different schools of thought. In addition to the varying ways in which psychotherapy has been conceptualized, there also exists a long-standing gap between psychotherapy research and how it is conducted in actual clinical practice. Finally, there also exists a tendency to place great emphasis on what is new, often rediscovering or reinventing past contributions. This article describes each of these impediments to obtaining consensus and offers some suggestions for what might be done to address them.

Here is an excerpt:

There are at least three problematic issues that seem to contribute to the difficulty we have in obtaining a consensus within the field of psychotherapy: The first involves our long-standing practice of solely working within theoretical orientations or eclectic combinations of orientations. Moreover, not agreeing with those having other frameworks on how to bring about therapeutic change results in the proliferation of schools of therapy (Goldfried, 1980). The second issue involves the longstanding gap between research and practice, where many therapists may fail to see the relevance to their day-to-day clinical practice and also where many researchers do not make systematic use of clinical observations as a means of guiding their research (Goldfried, 1982).2 The third issue is our tendency to neglect past contributions to the field (Goldfried, 2000). We do not build on our previous body of knowledge but rather rediscover what we already know or—even worse—ignore past work and replace it with something new. What follows is a description of how these three issues prevent psychotherapy from achieving a consensus, after which there will be a consideration of some possible steps that might be taken in working toward a resolution of these issues.

The article is here, behind a paywall.

Friday, October 26, 2018

Ethics, a Psychological Perspective

Andrea Dobson
www.infoq.com
Originally posted September 22, 2018

Key Takeaways
  • With emerging technologies like machine learning, developers can now achieve much more than ever before. But this new power has a down side. 
  • When we talk about ethics - the principles that govern a person's behaviour - it is impossible to not talk about psychology. 
  • Processes like obedience, conformity, moral disengagement, cognitive dissonance and moral amnesia all reveal why, though we see ourselves as inherently good, in certain circumstances we are likely to behave badly.
  • Recognising that although people aren’t rational, they are to a large degree predictable, has profound implications on how tech and business leaders can approach making their organisations more ethical.
  • The strongest way to make a company more ethical is to start with the individual. Companies become ethical one person at a time, one decision at a time. We all want to be seen as good people, known as our moral identity, which comes with the responsibility to have to act like it.