Essec\Faculty\Model\Contribution {#2233
#_index: "academ_contributions"
#_id: "16055"
#_source: array:26 [
"id" => "16055"
"slug" => "16055-moral-deliberation-reduces-peoples-intentions-to-share-headlines-they-recognize-as-fake-news"
"yearMonth" => "2025-11"
"year" => "2025"
"title" => "Moral Deliberation Reduces People’s Intentions to Share Headlines They Recognize as "Fake News""
"description" => "EFFRON, D., QIU, J. et SHULMAN, D. (2025). Moral Deliberation Reduces People’s Intentions to Share Headlines They Recognize as "Fake News". <i>Journal of Experimental Psychology: general</i>, Forthcoming."
"authors" => array:3 [
0 => array:3 [
"name" => "QIU Judy"
"bid" => "B00813631"
"slug" => "qiu-judy"
]
1 => array:1 [
"name" => "EFFRON Daniel"
]
2 => array:1 [
"name" => "SHULMAN Deborah"
]
]
"ouvrage" => ""
"keywords" => []
"updatedAt" => "2025-10-29 15:36:35"
"publicationUrl" => null
"publicationInfo" => array:3 [
"pages" => ""
"volume" => "Forthcoming"
"number" => ""
]
"type" => array:2 [
"fr" => "Articles"
"en" => "Journal articles"
]
"support_type" => array:2 [
"fr" => "Revue scientifique"
"en" => "Scientific journal"
]
"countries" => array:2 [
"fr" => null
"en" => null
]
"abstract" => array:2 [
"fr" => """
Ranked among the most serious global threats, misinformation spreads in part because people share it on social media. Based on theories that people usually share misinformation unintentionally, interventions typically aim to curb misinformation’s spread by helping people distinguish fact from falsehood. However, people sometimes intentionally spread misinformation despite recognizing its falsity. Understanding and curbing this type of sharing requires new theory and tools. Leveraging insights from moral psychology, the present research examines whether people will be more reluctant to share misinformation when they think carefully about its moral implications. Engaging in such moral deliberation, we theorize, leads people to judge misinformation as more unethical to share, which inhibits them from forming intentions to share it. Five experiments (four pre-registered, N=2,509 U.S. and U.K. social-media users, including a demographically representative U.S. sample) tested\n
a moral-deliberation procedure in which participants list reasons why it would be ethical or unethical to share different news headlines on social media. This procedure––relative to control conditions that prompted non-moral deliberation, prompted non-deliberative thinking about morality, or included no prompt––reduced intentions to share fake news about business, health, and politics, even when the news was flagged as false. These effects were (a) larger when the fake news aligned with participants’ politics, (b) reversed for real news, (c) still observed after a delay, and (d) mediated by moral judgments. Our results offer a theoretical foundation for new tools to fight society’s “infodemic” of misinformation.
"""
"en" => """
Ranked among the most serious global threats, misinformation spreads in part because people share it on social media. Based on theories that people usually share misinformation unintentionally, interventions typically aim to curb misinformation’s spread by helping people distinguish fact from falsehood. However, people sometimes intentionally spread misinformation despite recognizing its falsity. Understanding and curbing this type of sharing requires new theory and tools. Leveraging insights from moral psychology, the present research examines whether people will be more reluctant to share misinformation when they think carefully about its moral implications. Engaging in such moral deliberation, we theorize, leads people to judge misinformation as more unethical to share, which inhibits them from forming intentions to share it. Five experiments (four pre-registered, N=2,509 U.S. and U.K. social-media users, including a demographically representative U.S. sample) tested\n
a moral-deliberation procedure in which participants list reasons why it would be ethical or unethical to share different news headlines on social media. This procedure––relative to control conditions that prompted non-moral deliberation, prompted non-deliberative thinking about morality, or included no prompt––reduced intentions to share fake news about business, health, and politics, even when the news was flagged as false. These effects were (a) larger when the fake news aligned with participants’ politics, (b) reversed for real news, (c) still observed after a delay, and (d) mediated by moral judgments. Our results offer a theoretical foundation for new tools to fight society’s “infodemic” of misinformation.
"""
]
"authors_fields" => array:2 [
"fr" => "Management"
"en" => "Management"
]
"indexedAt" => "2025-12-06T05:21:43.000Z"
"docTitle" => "Moral Deliberation Reduces People’s Intentions to Share Headlines They Recognize as "Fake News""
"docSurtitle" => "Articles"
"authorNames" => "<a href="/cv/qiu-judy">QIU Judy</a>, EFFRON Daniel, SHULMAN Deborah"
"docDescription" => "<span class="document-property-authors">QIU Judy, EFFRON Daniel, SHULMAN Deborah</span><br><span class="document-property-authors_fields">Management</span> | <span class="document-property-year">2025</span>"
"keywordList" => ""
"docPreview" => "<b>Moral Deliberation Reduces People’s Intentions to Share Headlines They Recognize as "Fake News"</b><br><span>2025-11 | Articles </span>"
"docType" => "research"
"publicationLink" => "<a href="#" target="_blank">Moral Deliberation Reduces People’s Intentions to Share Headlines They Recognize as "Fake News"</a>"
]
+lang: "fr"
+"_score": 8.714403
+"_ignored": array:2 [
0 => "abstract.en.keyword"
1 => "abstract.fr.keyword"
]
+"parent": null
}