Welcome to the Nexus of Ethics, Psychology, Morality, Philosophy and Health Care

Welcome to the nexus of ethics, psychology, morality, technology, health care, and philosophy

Thursday, April 5, 2018

Would You Opt for Immortality?

Michael Shermer
Quillette
Originally posted March 2, 2018

Here is an excerpt:

The idea of living forever, in fact, is not such a radical idea when you consider the fact that the vast majority of people already believe that they will do so in the next life. Since the late 1990s Gallup has consistently found that between 72 and 83 percent of Americans believe in heaven. Globally, rates of belief in heaven in other countries typically lag behind those found in America, but they are nonetheless robust. A 2011 Ipsos/Reuters poll, for example, found that of 18,829 people surveyed across 23 countries,2 51 percent said they were convinced that an afterlife exists, ranging from a high of 62 percent of Indonesians and 52 percent of South Africans and Turks, to a low of 28 percent of Brazilians and only 3 percent of the very secular Swedes.

So powerful and pervasive are such convictions that even a third of agnostics and atheists proclaim belief in an afterlife. Say what? A 2014 survey conducted by the Austin Institute for the Study of Family and Culture on 15,738 Americans between the ages of 18 and 60 found that 13.2 percent identify as atheist or agnostic, and 32 percent of those answered in the affirmative to the question: “Do you think there is life, or some sort of conscious existence, after death?”

Depending on what these people believe about what, exactly, is resurrected in the next life—just your soul, or both your body and your soul—the belief among religious people that “you” will continue indefinitely in some form in the hereafter is not so different in principle from what the scientific immortalists are trying to accomplish in the here and now.

The article is here.

Moral Injury and Religiosity in US Veterans With Posttraumatic Stress Disorder Symptoms

Harold Koenig and others
The Journal of Nervous and Mental Disease: February 28, 2018

Abstract

Moral injury (MI) involves feelings of shame, grief, meaninglessness, and remorse from having violated core moral beliefs related to traumatic experiences. This multisite cross-sectional study examined the association between religious involvement (RI) and MI symptoms, mediators of the relationship, and the modifying effects of posttraumatic stress disorder (PTSD) severity in 373 US veterans with PTSD symptoms who served in a combat theater. Assessed were demographic, military, religious, physical, social, behavioral, and psychological characteristics using standard measures of RI, MI symptoms, PTSD, depression, and anxiety. MI was widespread, with over 90% reporting high levels of at least one MI symptom and the majority reporting at least five symptoms or more. In the overall sample, religiosity was inversely related to MI in bivariate analyses (r = −0.25, p < 0.0001) and multivariate analyses (B = −0.40, p = 0.001); however, this relationship was present only among veterans with severe PTSD (B = −0.65, p = 0.0003). These findings have relevance for the care of veterans with PTSD.

The paper is here.

Wednesday, April 4, 2018

Musk and Zuckerberg are fighting over whether we rule technology—or it rules us

Michael Coren
Quartz.com
Originally posted April 1, 2018

Here is an excerpt:

Musk wants to rein in AI, which he calls “a fundamental risk to the existence of human civilization.” Zuckerberg has dismissed such views calling their proponents “naysayers.” During a Facebook live stream last July, he added, “In some ways I actually think it is pretty irresponsible.” Musk was quick to retort on Twitter. “I’ve talked to Mark about this,” he wrote. “His understanding of the subject is limited.”

Both men’s views on the risks and rewards of technology are embodied in their respective companies. Zuckerberg has famously embraced the motto “Move fast and break things.” That served Facebook well as it exploded from a college campus experiment in 2004 to an aggregator of the internet for more than 2 billion users.

Facebook has treated the world as an infinite experiment, a game of low-stakes, high-volume tests that reliably generate profits, if not always progress. Zuckerberg’s main concern has been to deliver the fruits of digital technology to as many people as possible, as soon as possible. “I have pretty strong opinions on this,” Zuckerberg has said. “I am optimistic. I think you can build things and the world gets better.”

The information is here.

Simple moral code supports cooperation

Charles Efferson & Ernst Fehr
Nature
Originally posted March 7, 2018

The evolution of cooperation hinges on the benefits of cooperation being shared among those who cooperate. In a paper in Nature, Santos et al. investigate the evolution of cooperation using computer-based modelling analyses, and they identify a rule for moral judgements that provides an especially powerful system to drive cooperation.

Cooperation can be defined as a behaviour that is costly to the individual providing help, but which provides a greater overall societal benefit. For example, if Angela has a sandwich that is of greater value to Emmanuel than to her, Angela can increase total societal welfare by giving her sandwich to Emmanuel. This requires sacrifice on her part if she likes sandwiches. Reciprocity offers a way for benefactors to avoid helping uncooperative individuals in such situations. If Angela knows Emmanuel is cooperative because she and Emmanuel have interacted before, her reciprocity is direct. If she has heard from others that Emmanuel is a cooperative person, her reciprocity is indirect — a mechanism of particular relevance to human societies.

A strategy is a rule that a donor uses to decide whether or not to cooperate, and the evolution of reciprocal strategies that support cooperation depends crucially on the amount of information that individuals process. Santos and colleagues develop a model to assess the evolution of cooperation through indirect reciprocity. The individuals in their model can consider a relatively large amount of information compared with that used in previous studies.

The review is here.

Tuesday, April 3, 2018

Cambridge Analytica: You Can Have My Money but Not My Vote

Emily Feng-Gu
Practical Ethics
Originally posted March 31, 2018

Here is an excerpt:

On one level, the Cambridge Analytica scandal concerns data protection, privacy, and informed consent. The data involved was not, as Facebook insisted, obtained via a ‘breach’ or a ‘leak’. User data was as safe as it had always been – which is to say, not very safe at all. At the time, the harvesting of data, including that of unconsenting Facebook friends, by third-party apps was routine policy for Facebook, provided it was used only for academic purposes. Cambridge researcher and creator of the third-party app in question, Aleksandr Kogan, violated the agreement only when the data was passed onto Cambridge Analytica. Facebook failed to protect its users’ data privacy, that much is clear.

But are risks like these transparent to users? There is a serious concern about informed consent in a digital age. Most people are unlikely to have the expertise necessary to fully understand what it means to use online and other digital services.  Consider Facebook: users sign up for an ostensibly free social media service. Facebook did not, however, accrue billions in revenue by offering a service for nothing in return; they profit from having access to large amounts of personal data. It is doubtful that the costs to personal and data privacy are made clear to users, some of which are children or adolescents. For most people, the concept of big data is likely to be nebulous at best. What does it matter if someone has access to which Pages we have Liked? What exactly does it mean for third-party apps to be given access to data? When signing up to Facebook, I hazard that few people imagined clicking ‘I agree’ could play a role in attempts to influence election outcomes. A jargon laden ‘terms and conditions’ segment is not enough to inform users regarding what precisely it is they are consenting to.

The blog post is here.

AI Has a Hallucination Problem That's Proving Tough to Fix

Tom Simonite
wired.com
Originally posted March 9, 2018

Tech companies are rushing to infuse everything with artificial intelligence, driven by big leaps in the power of machine learning software. But the deep-neural-network software fueling the excitement has a troubling weakness: Making subtle changes to images, text, or audio can fool these systems into perceiving things that aren’t there.

That could be a big problem for products dependent on machine learning, particularly for vision, such as self-driving cars. Leading researchers are trying to develop defenses against such attacks—but that’s proving to be a challenge.

Case in point: In January, a leading machine-learning conference announced that it had selected 11 new papers to be presented in April that propose ways to defend or detect such adversarial attacks. Just three days later, first-year MIT grad student Anish Athalye threw up a webpage claiming to have “broken” seven of the new papers, including from boldface institutions such as Google, Amazon, and Stanford. “A creative attacker can still get around all these defenses,” says Athalye. He worked on the project with Nicholas Carlini and David Wagner, a grad student and professor, respectively, at Berkeley.

That project has led to some academic back-and-forth over certain details of the trio’s claims. But there’s little dispute about one message of the findings: It’s not clear how to protect the deep neural networks fueling innovations in consumer gadgets and automated driving from sabotage by hallucination. “All these systems are vulnerable,” says Battista Biggio, an assistant professor at the University of Cagliari, Italy, who has pondered machine learning security for about a decade, and wasn’t involved in the study. “The machine learning community is lacking a methodological approach to evaluate security.”

The article is here.

Monday, April 2, 2018

Ethics and sport have long been strangers to one another

Kenan Malik
The Guardian
Originally posted March 8, 2018

Here is an excerpt:

Today’s great ethical debate is not about payment but drugs. Last week, the digital, culture, media and sport select committee accused Bradley Wiggins of “crossing the ethical line” for allegedly misusing drugs allowed for medical purposes to enhance performance.

The ethical lines over drug use are, however, as arbitrary and irrational as earlier ones about payment. Drugs are said to be “unnatural” and to provide athletes with an “unfair advantage”. But virtually everything an athlete does, from high-altitude training to high-protein dieting, is unnatural and seeks to gain an advantage.

EPO is a naturally produced hormone that stimulates red blood cell production, so helping endurance athletes. Injections of EPO are banned in sport. Yet Chris Froome is permitted to sleep in a hypoxic chamber, which reduces oxygen in the air, forcing his body to produce more red blood cells. It has the same effect as EPO, is equally unnatural and provides an advantage. Why is one banned but not the other?

The article is here.

The Grim Conclusions of the Largest-Ever Study of Fake News

Robinson Meyer
The Atlantic
Originally posted March 8, 2018

Here is an excerpt:

“It seems to be pretty clear [from our study] that false information outperforms true information,” said Soroush Vosoughi, a data scientist at MIT who has studied fake news since 2013 and who led this study. “And that is not just because of bots. It might have something to do with human nature.”

The study has already prompted alarm from social scientists. “We must redesign our information ecosystem for the 21st century,” write a group of 16 political scientists and legal scholars in an essay also published Thursday in Science. They call for a new drive of interdisciplinary research “to reduce the spread of fake news and to address the underlying pathologies it has revealed.”

“How can we create a news ecosystem … that values and promotes truth?” they ask.

The new study suggests that it will not be easy. Though Vosoughi and his colleagues only focus on Twitter—the study was conducted using exclusive data which the company made available to MIT—their work has implications for Facebook, YouTube, and every major social network. Any platform that regularly amplifies engaging or provocative content runs the risk of amplifying fake news along with it.

The article is here.

Sunday, April 1, 2018

Sudden-Death Aversion: Avoiding Superior Options Because They Feel Riskier

Jesse Walker, Jane L. Risen, Thomas Gilovich, and Richard Thaler
Journal of Personality and Social Psychology, in press

Abstract

We present evidence of Sudden-Death Aversion (SDA) – the tendency to avoid “fast” strategies that provide a greater chance of success, but include the possibility of immediate defeat, in favor of “slow” strategies that reduce the possibility of losing quickly, but have lower odds of ultimate success. Using a combination of archival analyses and controlled experiments, we explore the psychology behind SDA. First, we provide evidence for SDA and its cost to decision makers by tabulating how often NFL teams send games into overtime by kicking an extra point rather than going for the 2-point conversion (Study 1) and how often NBA teams attempt potentially game-tying 2-point shots rather than potentially game-winning 3-pointers (Study 2). To confirm that SDA is not limited to sports, we demonstrate SDA in a military scenario (Study 3). We then explore two mechanisms that contribute to SDA: myopic loss aversion and concerns about “tempting fate.” Studies 4 and 5 show that SDA is due, in part, to myopic loss aversion, such that decision makers narrow the decision frame, paying attention to the prospect of immediate loss with the “fast” strategy, but not the downstream consequences of the “slow” strategy. Study 6 finds people are more pessimistic about a risky strategy that needn’t be pursued (opting for sudden death) than the same strategy that must be pursued. We end by discussing how these twin mechanisms lead to differential expectations of blame from the self and others, and how SDA influences decisions in several different walks of life.

The research is here.