Abstracts

advertisement
List of Abstracts (ordered alphabetically by first author)
Expectations and Reflection Explain the Knobe Effect
Mark Alfano, Paul Stey & Brian Robinson
Building on the work of Alfano, Beebe, & Robinson, we argue that the key to
understanding the Knobe effect is not the moral or prudential valence of the
consequences of the protagonist’s behavior but expectation (in)congruence. Our
thesis breaks into three parts: a description of the psychological conditions
under which the effect can be observed, a link between these conditions and
norm (in)congruence, and a rational reconstruction of the conditions and the
link.
At the most superficial level of analysis, we claim that people are more
inclined to attribute a wide variety of mental attitudes to an actor who produces
an effect contrary to expectations. This claim is supported by a new experimental
finding that expectations mediate mental state attributions in Knobe effect
vignettes. In all Knobe effect studies to date, participants answer questions only
after they find out what the protagonist does. This makes it impossible to probe
what they expect him to do before he acts. Unlike any previous study, our
experiment asked participants to state what they expected the protagonist to do
before they learned what he decided. Statistical analysis reveals that these
expectations explain most of the variance in subsequent attributions of mental
attitudes. When the protagonist acts contrary to expectations, participants say he
does so intentionally, but not otherwise.
A deeper level of analysis reveals that expectations are influenced by
salient norms, so when someone violates a salient norm (be it moral, prudential,
legal, aesthetic, conventional, or even merely descriptive), he typically acts
contrary to expectations and thus induces higher levels of mental state
attribution. This influence of norms on expectations explains why many
interpreters of the Knobe effect have been tempted to link it to morality.
Violating a moral norm is one way of going against expectations, but there are
many others, some of which involve adhering to a moral norm while violating
some competing norm. Moreover, the link is crucially between expectations and
salient norms, not just expectations and norms punkt. We illustrate this point by
varying our vignettes. In all conditions, the protagonist is choosing whether to
invest his inheritance in a retirement savings account or donate it to a worthy
charity. Investing conforms to a prudential norm while violating a moral
injunction to help the needy; donating conforms to the moral norm while
violating a principle of prudence. Thus, in all of our conditions, two norms are
relevant, and only one can be satisfied. However, they are not always both
salient. In different conditions, an interlocutor raises neither, one, or both of
these norms to salience, which in turn influences both expectations and mental
state attributions. Participants are more inclined to say that the protagonist
intentionally does not help when the helping norm is salient; they are also more
inclined to say that the protagonist intentionally does not prepare for retirement
when the prudential norm is salient.
The ultimate level of analysis appeals to the rationality of both forming
and attributing mental states in the way just described. It makes sense to pause
and deliberate when the action you’re about to take would violate a salient norm.
1
Deliberation in turn leads to the formation of various mental attitudes, such as
beliefs, desires, and intentions. Since we have limited cognitive powers, it makes
sense to curtail our deliberative engagement to those cases where ignorance
would be most deleterious. Such cases typically involve the violation of norms, so
it would be reasonable to deliberate more about potential norm-violation than
about potential norm-conformity. Knobe effect studies in the literature typically
ask participants to attribute mental states (intention, belief, desire, etc.). Our
experiment branches out to the attribution of mental processing. We asked
participants not only whether they thought the protagonist intentionally brought
about certain effects but also to what extent he considered the potential
consequences of his behavior. Our findings in this regard corroborate our
interpretation: participants attribute high levels of deliberation if and only if the
agent violates a salient norm.
Probabilistic Inference in the Trolley Problem and its Variants
Chris Brand & Mike Oaksford
One of the most important developments within the psychology of reasoning and
decision making over the past two decades has been the finding that people do
not reason as if the premises of an argument were certain. Instead, it has been
demonstrated that people reason in a probabilistic fashion. Curiously, this has
received little attention within moral psychological research, despite it having
profound implications for the field.
As an example, within standard interpretations the trolley problem
generally elicits a utilitarian choice and the footbridge problem elicits a
deontological one; within a probabilistic framework, both problems might be
producing a utilitarian response. In the case of the trolley problem, diverting the
train to a different track would render the death of the five on the original track
as very unlikely; furthermore, the death of the man on the second track is by no
means certain, as Foot herself noted in the original proposal of the dilemma.
Pulling the lever is therefore likely to lead to a very good outcome, in which at
least five of the men do not die.
In the footbridge problem, however, pushing someone in front of a train is
very likely to lead to their death, but given what most people know about trains
may also be seen as very unlikely to prevent the death of the five. Pushing
someone in front of the train would therefore be expected to lead to a worse
possible outcome than doing nothing, and should not be condoned by a good
utilitarian. That probabilistic utilitarian reasoning can lead to the standard
responses observed in experiments in ethics is easily shown; what has not yet
been demonstrated is that people do make such probabilistic inferences when
evaluating moral scenarios.
Although some researchers have previously discussed the potential
relationship of probabilistic judgements to moral reasoning, they have generally
either dismissed such effects as covariants (such as Greene et al., 2009) or failed
to experimentally test the possibility (Sloman et al., 2009). This seems like a
curious omission. As such, we will report on the preliminary findings of a
currently-ongoing study investigating whether there is a significant relationship
between the judged permissibility of a number of trolley problems, and the
estimated likelihood of various possible outcomes within them.
2
A possible solution to the trolley problem: justice towards groups and the
timing of the attribution of individual positions
Florian Cova & David Spector
We propose a novel explanation to the well-known trolley problem. The existing
research has shown that, when faced with a hypothetical dilemma involving the
sacrifice of one life in order to save five, respondents tend to provide very
different solutions depending on the framing of the scenario (Hauser et al.,
2007).
We tested on over 600 participants two variants of both the standard
scenarios and variations upon these scenarios (using both original and already
existing cases). In the first variant, all individuals originally belong to the same
group and their final positions were the result of late decisions. In the other
variant, the potentially sacrificed person and the potentially saved five belong to
different groups from the outset. We find that (i) respondents were more willing
to sacrifice one in order to save five if all individuals originally belonged to the
same group, and (ii) when not told anything about group membership,
respondents made implicit assumptions, which differed across scenarios. Also,
(iii) our results allowed us to rule out the traditional account in terms of the
doctrine of double effect (e.g., Mikhail, 2007).
On the basis of these results, we argue that the apparent incoherence of
answers to the dilemma in different but equivalent scenarios results from the
combination of diverging implicit assumptions about group membership and an
aversion for inter-group redistribution. We conjecture that this aversion results
from the combination of a mild omission bias, rawlsian preferences, and
reasoning as if decisions were made behind a veil of ignorance that is pierced
after group membership is determined but before within-group positions are.
Additionally, we argue that the same factors can account for problems of
replication in the trolley literature, in particular about the famous “man-on-theloop” case (compare for example Hauser et al., 2007 with Greene et al., 2009).
The Philosopher in the Theater
Fiery Cushman
Where do moral principles come from? Some moral philosophers derive
principles by reflecting on their own intuitions. We propose that ordinary people
do much the same thing. Based on a several case-studies, we suggests that
people generalize explicit moral principles from automatic moral intuitions,
codifying reliable consistencies in their 'gut feelings'. Explicit moral principles
therefore reflect the basic structure of the cognitive systems that generate our
intuitive moral judgments. And, because intuitive moral judgments depend on
an assessment of causal responsibility and mental culpability, those same causal
and mental state analyses figure prominently in explicit moral
theories. Interestingly, certain psychological 'quirks' of reasoning about
causation and mental states that show up in our moral judgments therefore also
show up in our moral principles. In this sense, our moral principles reflect not
just facts about the world, but also peculiar structures of our minds. While the
"Cartesian theater" has sometimes been mocked as a psychological model, we
propose it as a useful analogy. There is a philosopher in the theater, trying to
3
make principled sense of her own moral intuitions -- among other psychological
systems -- putting on the show.
On the attribution of externalities
Urs Fischbacher
Do people blame or praise others for producing negative or positive
externalities? The experimental philosopher Knobe conducted a questionnaire
study that revealed that people blame others for foreseen negative externalities
but do not praise them for foreseen positive ones. We find that the major
determinant of the Knobe effect is the relative distribution of economic power
among the agents. We confirm the Knobe effect only in situations where the
producer of the externality holds the higher economic status and the positive
externalities are small. Switching economic power makes the Knobe effect
vanish. The Knobe effect is even reversed in settings with large positive
externalities. Our results are in line with theoretical predictions by Levine.
Philosophical Dilemmas, Philosophical Personality, and Philosophy in Action
Adam Feltz
Perhaps personality traits substantially influence one's philosophically relevant
intuitions. This suggestion is not only possible, it is consistent with a growing
body of empirical research: Personality traits have been shown to be
systematically related to diverse intuitions concerning some fundamental
philosophical debates. This fact, in conjunction with the plausible principle that
almost all adequate philosophical views should take into account all available and
relevant evidence, calls into question some prominent approaches to traditional
philosophical projects. I explain how the growing body of evidence challenging
some of the uses of intuitions in philosophy, and I defend this challenge from some
criticisms of empirically based worries about intuitions in philosophy. I conclude
by suggesting two possibly profound implications. First, some dominant
traditional philosophical projects must become substantially more empirically
oriented. Second, much of philosophy ought to become substantially more applied.
Your Money or Your Life: Varying the outcomes in trolley problems
Natalie Gold, Briony Pulford & Andrew Colman
Trolley problems are used in order to probe intuitions about the morality of
harming one person in order to prevent a greater harm to others. Philosophers
use the moral intuitions derived from them to achieve a reflective equilibrium
between intuitions and philosophical principles, psychologists to investigate
processes of moral judgment and moral decision-making. Trolley problems
involve trade-offs between life and death, they are unrealistic scenarios, and they
are hypothetical scenarios. Do the results from trolley problems generalize to
harms other than death, to more familiar scenarios, and to judgments about real
events? We present three experiments designed to examine these questions. In
the first experiment, we present evidence that the difference in moral intuitions
4
between bystander and footbridge scenarios is replicated across different
domains and levels of physical and non-physical harm. In a second experiment,
we transplant the trolley problem to a more familiar scenario (also involving
financial harms), and discover that judgments are different depending on
whether the agent is a “bystander”, or onlooker, or a “passenger”, who is already
involved in the scenario, but that changing of context reverses the direction of
the effect. In the third experiment, participants make judgments about the
morality of trade-offs between small financial harms that are actually happening
in the lab as they make their judgments, enabling us to investigate bystanderfootbridge, actor-observer, and order effects in an incentive compatible
scenarios.
Emotional Arousal and Moral Judgment
Zachary Horne, Derek Powell
Moral psychological research has revealed that both reasoning and emotion play
a role in moral decision-making. A debate persists, however, about which of
these two processes is primary, and in what contexts. Strictly speaking, evidence
that emotion plays a role in moral decision-making is not evidence against the
role of reasoning, and vice versa. This study seeks to adjudicate the debate by
examining the causal relationship between people’s emotional states and their
moral judgments. Using self-report data from a commonly used emotion
measure (PANAS-X), we measured participants’ emotional responses to several
moral dilemmas and examined whether or not their emotional states could
predict their moral judgments. Compared to non-moral control dilemmas, our
findings indicate that emotions are reliably cued during moral scenarios. Despite
the intense emotions participants experienced, their emotional reactions were
not predictive of their moral judgments. Our findings, in conjunction with related
findings, call into question models of moral cognition that seek to explain
people's behavior in moral dilemmas entirely in terms of emotional arousal.
Who Makes A Tough Call?: An Emotion Regulation Approach To Moral
Decision-Making
Joo-A Julia Lee
Moral dilemmas often invoke strong emotional reactions. McClure et al. (2007)
argued that detecting and regulating competition between cognitive and
emotional processes is associated with activity in a dorsal region of the anterior
cingulate cortex (ACC). Their model suggests that our brain may be involved in
constantly monitoring the conflicts between the two processes, and potentially
overriding our emotional reactions by regulating and controlling them. Drawing
on their hypothesis on conflict monitoring, we contribute to this theory of
conflict monitoring by showing that emotion regulation plays a key role in
making utilitarian decisions by inhibiting the influence of negative emotions that
are associated with the prospect of harm.
We first predicted that these emotional reactions affect our decisions in a
critical way by interfering the deliberate, utilitarian decision-making process.
More specifically, we hypothesized that individuals who are told to regulate their
5
emotions would be more likely to make utilitarian decisions, compared to those
who are not asked to suppress their emotions. We also expected that the
relationship between one’s emotion regulation strategy and moral decisions
would be qualified by the extent of physiological arousal when considering the
moral dilemma such that the relationship between emotion regulation and
utilitarian decisions would be stronger among those who felt strong emotions.
Lastly, we hypothesized that people who regulate their emotions will
have heightened moral clarity perceptions in unrelated moral dilemma
situations, compared to those who do not regulate their emotions. We also
predicted that this relationship between emotion regulation and moral clarity
would be mediated by one’s revealed preference for utilitarian choice, thus
changing one’s moral judgment as well.
We tested our main hypotheses in two studies. In Study 1, we use a
correlational design and examine whether individual differences in emotion
regulation are correlated with one’s decisions in solving moral dilemmas. In
Study 2, we directly test the causal relationship between emotion regulation
strategies (in particular, suppression) and one’s preference for utilitarian
decisions in moral dilemmas, as well as one’s moral clarity judgment.
The two studies provided a critical link between regulating emotions and
moral judgment and decision-making in the dilemma situation. Not only
participants made a more utilitarian decision when they were told to suppress
strong emotional reactions they experienced, but their utilitarian preference also
carried over to increase their perception of moral clarity. In turn, this result
indicates that strong emotional arousal during the video made participants more
averse to the utilitarian option, and this aversion also led to the lack of moral
clarity, by perceiving the ethical dilemmas as more ambiguous.
Rational learners and non-utilitarian rules
Shaun Nichols
Hundreds of studies on moral dilemmas show that people’s judgments do not
conform to utilitarian principles. However, the exact nature of this
nonconformity remains unclear. Some maintain that people rely on deontological
“side constraints” that are insensitive to cost-benefit analysis. However, the
scenarios that are used to support this intuition, e.g., the magistrate and the mob,
contain an important confound. In these cases, we consider whether it is
appropriate for one person to violate a moral rule in order to prevent others
from committing similar violations. In that case, people tend to say that it would
be wrong to violate the rule. In a series of experiments, we showed that people
give very different responses when the question is whether an agent should
violate a moral rule so that she herself doesn’t have to commit more such
violations in the future. This suggests that a critical feature of our moral rules is
that they function in an intra-agent, rather than inter-agent manner. But this
raises a further question – why do our rules have this non-utilitarian character?
One prominent view (e.g. Mikhail 2007) holds that the structure of moral rules
plausibly depends on an innate moral grammar. We propose instead that given
the evidence that the young child has, a rational Bayesian learner would in fact
arrive at non-utilitarian rules.
6
Investigating the Effects of Framing in Trolley Problems
Briony Pulford, Andrew Colman, & Natalie Gold
In an examination of judgments in trolley problems we studied 15 scenarios with
slightly different framings to disentangle different effects that may be
occurring. Via an on-line survey 1,853 participants read one of the scenarios,
chose either Yes or No (if killing the one to save the five was morally
permissible), and rated attributes of the person who took the action, such as
whether they caused the one man to die or were to blame. Whether turning the
train is morally acceptable was also rated. We replicated the findings of Hauser,
Cushman, Young, Jin & Mikhail (2007) using the Ned, Oscar and Denise scenarios,
and found almost identical results for the percentage of people who agreed that
it was morally permissible to kill one person to save five. Possible confounds
with the scenarios such as the train driver fainting or not, and the gender of the
person in the scenario were found to be non-significant and were excluded as
causes. More importantly, referring to the man as a ‘heavy object’ or not did not
significantly influence people’s judgments. Bystander-passenger differences
were clear in the side-track scenarios but non-significant in the loop-track
scenarios.
The self/other incongruity in moral psychology: data and implications
Regina Rini
Recent work in experimental philosophy (e.g. Nadelhoffer and Feltz 2008)
suggests that moral judgments display a self/other incongruity (which is
sometimes also called the actor/observer bias). That is, moral judgments of
particular actions seem to depend in part upon whether the subject imagines
herself, or someone else, in the position of actor. This result does not seem to
have received the attention it merits. In this paper I argue that these findings
present a potentially unique challenge to the reliability of moral judgment. In my
view, the self/other incongruity is not simply another cognitive bias. Rather, it
challenges a fundamental presupposition of contemporary western moral
philosophy: that moral judgments display universalizability. If our moral
judgments routinely do not assess all similarly-situated agents equally, then by
failing the universalizability requirement they simply do not qualify as moral
judgments at all! I discuss several plausible responses to this worrying thesis,
but suggest that resolution must await further empirical details regarding the
cognitive processes underlying the incongruity. I also present findings from my
own experimental research, attempting to replicate and generalize the results
from Nadelhoffer and Feltz, and revealing a previously unreported interaction
between the self/other incongruity and gender.
Deadly Risk: Probability and Harm in Moral Judgment
Tor Tarantola
7
Recent research in moral psychology has focused on the processes of causal
reasoning, theory‐of-mind, and their integration in the formation of moral
judgments. However, relevant causal inputs have largely been investigated in
binary terms: either a consequence occurs or it does not, is caused by an action
or is not. The present research examines the role of harm probability
independent of theory‐of‐mind and consequence. How does an actor’s
probability of harming his potential victim, independent of his knowledge and
the ultimate consequence, affect a third‐party’s moral evaluation of his behavior?
Study 1 reports the results of an experimental survey (n=837) that shows that
punishment, but not wrongness, is attenuated by a reduction in this type of harm
probability. Study 2 reports the thematic analysis of four focus group interviews
that probed the extent to which this phenomenon was consciously
acknowledged. It also examines the process by which theory‐of‐mind and causal
considerations are integrated in deliberative moral reasoning. Implications for
theories of moral cognition are discussed, and a tentative model is proposed.
Psychopathic trait affects the moral choice but not the moral judgement:
experimental results of dilemmas resolution
Sébastien Tassy, Christine Deruelle, Julien Mancini, Samuel Leistedt, Bruno Wicker
Psychopathy is a personality disorder frequently associated with antisocial
behaviours. Although psychopathy increases the probability of immoral
behaviour, studies on the influence of psychopathy on decision making during
moral dilemma evaluation yielded contradictory results. Psychopathy either
increased the probability of utilitarian response or not. Here, we propose that
judgement (abstract evaluation) and choice of behaviour may be underlied by
two distinct cognitive processes and that psychopathy could alter specifically the
moral choice of behaviour, leading to intact moral judgement but immoral
behaviour. To explore this hypothesis, we tested the effect of the level of
psychopathic trait in a student sample on both judgment (”Is it acceptable to .... in
order to....”) and choice of behavior (”Would you ....in order to....?”) responses to
moral dilemmas. Results show that high psychopathic trait increases the
probability of utilitarian response for moral choice of behaviour but not for
moral judgement. The reason would be that psychopathy, which seems to alter
the network involving amygdala and the ventromedial prefrontal cortex, tends to
reduce the empathy for the victims, and would thus increase the acceptance to
make suffer someone to enhance the aggregate welfare. These results are in
favour of a dissociation of the cognitive processes that govern choice of
behaviour and judgement. It also explains the discrepancy of results of previous
studies on the effect of psychopathy on moral dilemmas resolution, which tested
indistinctively either moral judgement or moral choice of behaviour.
Mortality Salience and Morality: Thinking About Death Makes People Less
Utilitarian
Bastien Trémolière & Jean-François Bonnefon
8
The dual-process theory of moral judgment postulates that utilitarian responses
to moral dilemmas require to mobilize controlled processes that draw on limited
cognitive resources. In parallel, Terror Management Theory postulates that these
same resources are mobilized when one is reminded of one's own future
physical death (in order to suppress these thoughts out of focal attention). Crosspollinating these two perspectives, we predicted that people under mortality
salience (MS) would be less likely to give utilitarian responses to moral
dilemmas. Experiment 1 showed that the frequency of utilitarian responses was
lower when participants were presented with dilemma involving non-lethal
harm, than when they were presented with more traditional life-and-death
dilemma. Experiment 2 introduced an exogenous manipulation of MS (asking
participants to jot down some brief reactions to the idea of their future physical
death) and used non-lethal harm dilemma only. Participants under MS were less
likely to give utilitarian responses, compared to participants in a control group
who had to think about physical pain. Experiment 3 cross manipulated MS and
three degrees of cognitive load. Results replicated the effect of MS, ans showed
that only extra high load had a comparable effect of responses. The combination
of MS and extra-high load decreased the frequency of utilitarian responses by 40
percentage points. In addition to providing novel support to a dual- process
approach to moral judgment, these findings raise the worrying question of
whether mortality salience effects might shape private judgment and public
debates, by preventing a full reflective attention to the available arguments, since
these issues often involve matters of life and
When Can('t) We Trust Our Moral Intuitions in Distributive Cases?
Alex Voorhoeve, Ken Binmore and Brian Wallace
I examine the reliability of our moral intuitions in cases in which we must choose
to save either (1) a smaller number of people from greater harm, or, instead, (2)
a larger number of people from lesser harm. Such choices involve tradeoffs
between two dimensions (the number of people saved and the severity of harm
from which they are saved) for which it is difficult to specify the right "rate of
exchange."
There is evidence that when faced with non-moral choices that involve this kind
of difficult trade-off, a significant proportion of people will use a heuristic known
as "similarity-based decision-making." Roughly, this involves neglecting
dimensions along which alternatives are similar, and basing one's decisions on
only the dimension(s) along which they are dissimilar (Rubinstein 1988). The
use of this heuristic can lead to an under-weighting of dimensions along which
alternatives are similar. It may also lead to violations of principles of rational
choice. This means that, insofar as they use this heuristic in intuitive moral
choices, people's intuitions are suspect.
I report results of what is, to my knowledge, the first experiment on the use of
this heuristic in moral decisions. Eighty-two subjects were asked to make a
series of health care allocation choices of the kind outlined. The data reveals
strong evidence for the use of a similarity heuristic in a significant proportion
9
(around 30%) of subjects. It also reveals that these subjects generally
underweight similar dimensions, leading to violations of principles of rational
choice and to morally problematic choices.
I conclude that we should not trust our moral intuitions in cases which involve
difficult tradeoffs and in which alternatives are similar along one dimension.
Motivated Moral Perceivers
Jen Cole Wright, Michelle DiBartolo, Annie Galizio, & Evan Reinhold
People often make snap (“intuitive”) moral judgments, later highlighting
information that confirms their judgment (aka “motivated moral reasoning”).
Our studies examined whether people are also “motivated moral perceivers”,
disproportionally attending to visual information consistent with their preexisting moral judgments when evaluating morally challenging situations.
Participants were presented with dilemmas in which a person/people must die
for another person/people to live and then with visual images of the people
involved in the dilemmas. We predicted participants would show a visual
preference for the person/people they had decided to save and/or avoidance of
the person/people they decided to sacrifice.
In study 1, we used the ERICA eye-tracking system to track 266
participants’ eye movements as they read the classic Trolley or Footbridge case,
followed by images of the individual man and the five workers (counterbalanced
on screen side) in the case. After viewing the images, participants were asked
how willing (on a 7-point Likert scale) they were to pull the switch/push the
man off the bridge to save the five workers. We found that participants who
glanced first at the workers reported being more willing to pull the switch/push
the man off the bridge than those who glanced first at the man. Thus, our results
confirmed that participants’ visual attention patterns were predictive of which
person/people they had decided to kill/let die.
A limitation of this first study is that though we assumed participants
formed their judgment immediately upon reading the case (before seeing the
images), since we did not have participants report their judgments until
afterwards, there is no way to be sure. So, in study 2, we presented 115
participants with the Trolley case and had them answer the question before the
images. Once again, we found that participants whose first glance was at the
workers were more willing to pull the switch than those whose first glance was
at the individual.
These studies provide preliminary support for the hypothesis that people
display preference for visual information that is consistent with their moral
judgments. Having been confronted with a moral choice – namely, to kill an
individual in order to save a group – participants’ intuitive moral judgments (to
save or kill the individual) predicted where they first looked in both the
Trolley/Footbridge cases and in those Baby/Villager cases that were preceded
by the Trolley case. In those cases, participants looked first to the individual(s)
they intended to save/not kill.
10
Download