Misplaced Pages

Misinformation: Difference between revisions

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.
Browse history interactively← Previous editContent deleted Content addedVisualWikitext
Revision as of 12:31, 23 November 2020 editKeith D (talk | contribs)Autopatrolled, Administrators544,506 edits Fix cite date errorTag: AWB← Previous edit Latest revision as of 21:07, 31 December 2024 edit undoKMaster888 (talk | contribs)Extended confirmed users10,949 edits ce 
(691 intermediate revisions by more than 100 users not shown)
Line 1: Line 1:
{{Short description|Incorrect or misleading information}}
{{distinguish|Disinformation|Misinformation effect}} {{distinguish|Disinformation|Misinformation effect}}
] for the successful ] in the ]. The claim made by the sign was widely considered to have been an example of misinformation.<ref>{{cite web |last1=Henley |first1=Jon |title=Why Vote Leave's £350m weekly EU cost claim is wrong |url=https://www.theguardian.com/politics/reality-check/2016/may/23/does-the-eu-really-cost-the-uk-350m-a-week |website=The Guardian |date=10 June 2016 |access-date=14 October 2024}}</ref><ref>{{cite web |title=The UK's EU membership fee |url=https://fullfact.org/europe/our-eu-membership-fee-55-million/ |website=Full Fact |date=25 February 2016 |language=en |access-date=14 October 2024}}</ref><ref>{{cite web |title=Reality Check: Would Brexit mean extra £350m a week for NHS? |url=https://www.bbc.co.uk/news/uk-politics-eu-referendum-36040060 |website=BBC News |date=15 April 2016}}</ref><ref>{{cite web |last1=Ackrill |first1=Robert |title=Fact Check: how much does the UK actually pay to the EU? |url=https://theconversation.com/fact-check-how-much-does-the-uk-actually-pay-to-the-eu-58120 |website=The Conversation |date=27 April 2016}}</ref>]]
{{short description|False or incorrect information}}
'''Misinformation''' is false or inaccurate ] that is communicated regardless of an intention to deceive.<ref>{{Cite web|last=Merriam-Webster Dictionary|date=19 August 2020|title=Misinformation|url=https://www.merriam-webster.com/dictionary/misinformation|url-status=live|archive-url=|archive-date=|access-date=19 August 2020|website=}}</ref><ref>{{Cite web|last=Merriam-Webster Dictionary|date=19 August 2020|title=disinformation|url=https://www.merriam-webster.com/dictionary/disinformation|url-status=live|archive-url=|archive-date=|access-date=19 August 2020|website=Merriam-Webster}}</ref> Examples of misinformation are false rumors, insults, and pranks. ] is a species of misinformation that is deliberately deceptive, e. g. malicious ]es, ], and ].<ref>{{Cite journal|last1=Woolley|first1=Samuel C.|last2=Howard|first2=Philip N.|date=2016|title=Political Communication, Computational Propaganda, and Autonomous Agents|url=https://ijoc.org/index.php/ijoc/article/view/6298|journal=International Journal of Communication|volume=10|pages=4882–4890|via=|access-date=2019-10-22|archive-url=https://web.archive.org/web/20191022194727/https://ijoc.org/index.php/ijoc/article/view/6298|archive-date=2019-10-22|url-status=live}}</ref> The principal effect of misinformation is to elicit fear and suspicion among a population.<ref name=":0">{{Cite journal|last1=Chen|first1=Xinran|last2=Sin|first2=Sei-Ching Joanna|last3=Theng|first3=Yin-Leng|last4=Lee|first4=Chei Sian|date=September 2015|title=Why Students Share Misinformation on Social Media: Motivation, Gender, and Study-level Differences|journal=The Journal of Academic Librarianship|language=en|volume=41|issue=5|pages=583–592|doi=10.1016/j.acalib.2015.07.003}}</ref> News parody or satire can become misinformation if the unwary judge it to be credible and communicate it as if it were true. The words "misinformation" and "disinformation" have often been associated with the neologism "]", which some scholars define as "fabricated information that mimics news media content in form but not in organizational process or intent".<ref name="sfn">{{Cite journal|title=The science of fake news|journal = Science|volume = 359|issue = 6380|pages = 1094–1096|doi = 10.1126/science.aao2998|pmid = 29590025|year = 2018|last1 = Lazer|first1 = David M. J.|last2 = Baum|first2 = Matthew A.|last3 = Benkler|first3 = Yochai|last4 = Berinsky|first4 = Adam J.|last5 = Greenhill|first5 = Kelly M.|last6 = Menczer|first6 = Filippo|last7 = Metzger|first7 = Miriam J.|last8 = Nyhan|first8 = Brendan|last9 = Pennycook|first9 = Gordon|last10 = Rothschild|first10 = David|last11 = Schudson|first11 = Michael|last12 = Sloman|first12 = Steven A.|last13 = Sunstein|first13 = Cass R.|last14 = Thorson|first14 = Emily A.|last15 = Watts|first15 = Duncan J.|last16 = Zittrain|first16 = Jonathan L.|s2cid = 4410672|bibcode = 2018Sci...359.1094L}}</ref> '''Misinformation''' is incorrect or misleading ].<ref name=":2">{{Cite web|last=Merriam-Webster Dictionary|date=19 August 2020|title=Misinformation|url=https://www.merriam-webster.com/dictionary/misinformation|access-date=19 August 2020|archive-date=25 February 2019|archive-url=https://web.archive.org/web/20190225044837/https://www.merriam-webster.com/dictionary/misinformation|url-status=live}}</ref><ref>{{Cite journal |last=Fetzer |first=James H. |date=2004-05-01 |title=Information: Does it Have To Be True? |url=https://doi.org/10.1023/B:MIND.0000021682.61365.56 |journal=Minds and Machines |volume=14 |issue=2 |pages=223–229 |doi=10.1023/B:MIND.0000021682.61365.56 |s2cid=31906034 |issn=1572-8641}}</ref> Misinformation and ] are not interchangeable terms: Misinformation can exist with or without specific malicious intent whereas disinformation is distinct in that the information is ''deliberately'' deceptive and propagated.<ref>{{Cite journal |last1=Woolley |first1=Samuel C. |last2=Howard |first2=Philip N. |date=2016 |title=Political Communication, Computational Propaganda, and Autonomous Agents |url=https://ijoc.org/index.php/ijoc/article/view/6298 |url-status=live |journal=] |volume=10 |pages=4882–4890 |archive-url=https://web.archive.org/web/20191022194727/https://ijoc.org/index.php/ijoc/article/view/6298 |archive-date=2019-10-22 |access-date=2019-10-22}}</ref><ref name=":3">{{cite book |doi=10.1109/icict50521.2020.00076 |chapter=An Exploration of Disinformation as a Cybersecurity Threat |title=2020 3rd International Conference on Information and Computer Technologies (ICICT) |year=2020 |last1=Caramancion |first1=Kevin Matthe |pages=440–444 |isbn=978-1-7281-7283-5 |s2cid=218651389 }}</ref><ref>{{Cite web |last1=Karlova |first1=Natascha A. |last2=Fisher |first2=Karen E. |date=2013-03-15 |title=A social diffusion model of misinformation and disinformation for understanding human information behaviour |url=https://informationr.net/ir/18-1/paper573.html |access-date=2023-05-11 |website=informationr.net |archive-date=2023-05-11 |archive-url=https://web.archive.org/web/20230511204310/https://informationr.net/ir/18-1/paper573.html |url-status=live }}</ref><ref name=":42" /><ref>{{Cite journal |last=Diaz Ruiz |first=Carlos |date=2023-10-30 |title=Disinformation on digital media platforms: A market-shaping approach |journal=New Media & Society |doi=10.1177/14614448231207644 |issn=1461-4448|doi-access=free }}</ref> Misinformation can include inaccurate, incomplete, misleading, or false information as well as selective or half-truths. In January 2024, the ] identified misinformation and disinformation, propagated by both internal and external interests, to "widen societal and political divides" as the most severe global risks within the next two years.<ref>, World Economic Forum. {{ISBN|978-2-940631-64-3}}</ref>


Much research on how to correct misinformation has focused on ].<ref name=":11" /> However, this can be challenging because the ] does not necessarily apply well to beliefs in misinformation.<ref name=":4" /><ref name=":16" /> Various researchers have also investigated what makes people susceptible to misinformation.<ref name=":16"/> People may be more prone to believe misinformation because they are emotionally connected to what they are listening to or are reading. Social media has made information readily available to society at anytime, and it connects vast groups of people along with their information at one time.{{sfn|Aral|2020}} Advances in ] have impacted the way people communicate information and the way misinformation is spread.<ref name=":11">{{Cite journal |last1=Lewandowsky |first1=Stephan |last2=Ecker |first2=Ullrich K. H. |last3=Seifert |first3=Colleen M. |last4=Schwarz |first4=Norbert |last5=Cook |first5=John |date=2012 |title=Misinformation and Its Correction: Continued Influence and Successful Debiasing |journal=] |volume=13 |issue=3 |pages=106–131 |doi=10.1177/1529100612451018 |jstor=23484653 |pmid=26173286 |s2cid=42633 |doi-access=free }}</ref> Misinformation can influence people's beliefs about communities, politics, medicine, and more.{{sfn|Aral|2020}}<ref name="NYT-20240316" /> The term also has the potential to be used to obfuscate legitimate speech and warp political discourses.
==History==
The history of misinformation, along with that of ] and ], is part of the history of ].<ref name="icfj">{{Cite web|url=https://www.icfj.org/news/short-guide-history-fake-news-and-disinformation-new-icfj-learning-module|title=A short guide to the history of 'fake news' and disinformation|website=International Center for Journalists|access-date=2019-02-24|archive-url=https://web.archive.org/web/20190225103100/https://www.icfj.org/news/short-guide-history-fake-news-and-disinformation-new-icfj-learning-module|archive-date=2019-02-25|url-status=live}}</ref> Early examples cited in a 2017 article by Robert Darnton<ref>{{Cite web|url=https://www.nybooks.com/daily/2017/02/13/the-true-history-of-fake-news/|title=The True History of Fake News|website=The New York Review of Books|access-date=2019-02-24|date=2017-02-13|archive-url=https://web.archive.org/web/20190205222751/https://www.nybooks.com/daily/2017/02/13/the-true-history-of-fake-news/|archive-date=2019-02-05|url-status=live}}</ref> are the insults and smears spread among political rivals in Imperial and Renaissance Italy in the form of "]s". These are anonymous and witty verse named for the ] piazza and "]" in Rome, and in pre-revolutionary France as "canards", or printed broadsides that sometimes included an engraving to help convince readers to take their wild tales seriously.


The term came into wider recognition during the mid-1990s through the early 2020s, when its effects on public ideological influence began to be investigated. However, misinformation campaigns have existed for hundreds of years.<ref>{{Cite journal |last1=Bode |first1=Leticia |last2=Vraga |first2=Emily |date=23 June 2015 |title=In Related News, That was Wrong: The Correction of Misinformation Through Related Stories Functionality in Social Media |url=https://academic.oup.com/joc/article/65/4/619-638/4082315 |access-date=2024-01-31 |journal=Journal of Communication |volume=65 |issue=4 |pages=619–638 |doi=10.1111/jcom.12166 |s2cid=142769329 |archive-date=2022-11-12 |archive-url=https://web.archive.org/web/20221112201902/https://academic.oup.com/joc/article/65/4/619-638/4082315 |url-status=live }}</ref><ref>{{Cite web |last1=Posetti |first1=Julie |last2=Matthews |first2=Alice |date=June 23, 2018 |title=A Short Guide to the History of 'Fake News' and Disinformation: A New ICFJ Learning Module |url=https://www.icfj.org/news/short-guide-history-fake-news-and-disinformation-new-icfj-learning-module |access-date=2024-01-31 |website=International Center for Journalists |archive-date=2019-02-25 |archive-url=https://web.archive.org/web/20190225103100/https://www.icfj.org/news/short-guide-history-fake-news-and-disinformation-new-icfj-learning-module |url-status=live }}</ref>
The spread in Europe and North America of ]'s mechanized printing press increased the opportunities to spread English-language misinformation. In 1835, the New York Sun published the first large-scale news hoax, known as the "]". This was a series of six articles claiming to describe life on the Moon, "complete with illustrations of humanoid bat-creatures and bearded blue unicorns".<ref name=icfj /> The fast pace and sometimes strife-filled work of mass-producing news broadsheets also led to copies rife with careless factual errors and mistakes, such as the Chicago Tribune's infamous 1948 headline "]".
]
In the so-called ], ] have become a notable vector for the spread of misinformation, "fake news" and propaganda.<ref name="The Growing Role of Social Media in">{{cite journal |last1=Stawicki |first1=Stanislaw |last2=Firstenberg |first2=Michael |last3=Papadimos |first3=Thomas |title=The Growing Role of Social Media in International Health Security: The Good, the Bad, and the Ugly |journal=Global Health Security |volume=1 |issue=1 |pages=341–357}}</ref><ref name="sfn" /><ref name="vosoughi">{{cite journal |last1=Vosoughi |first1=Soroush |last2=Roy |first2=Deb |last3=Aral |first3=Sinan |s2cid=4549072 |title=The spread of true and false news online |journal=Science |volume=359 |issue=6380 |pages=1146–1151 |date=2018-03-09 |doi=10.1126/science.aap9559 |pmid=29590045 |bibcode=2018Sci...359.1146V |url=http://vermontcomplexsystems.org/share/papershredder/vosoughi2018a.pdf |access-date=2019-08-21 |archive-url=https://web.archive.org/web/20190429073158/http://vermontcomplexsystems.org/share/papershredder/vosoughi2018a.pdf |archive-date=2019-04-29 |url-status=dead }}</ref><ref name="tucker">{{cite journal |last1=Tucker |first1=Joshua A. |last2=Guess |first2=Andrew |last3=Barbera |first3=Pablo |last4=Vaccari |first4=Cristian |last5=Siegel |first5=Alexandra |last6=Sanovich |first6=Sergey |last7=Stukal |first7=Denis |last8=Nyhan |first8=Brendan |title=Social Media, Political Polarization, and Political Disinformation: A Review of the Scientific Literature |journal=Hewlett Foundation White Paper |url=https://hewlett.org/library/social-media-political-polarization-political-disinformation-review-scientific-literature/ |access-date=2019-03-05 |archive-url=https://web.archive.org/web/20190306111605/https://hewlett.org/library/social-media-political-polarization-political-disinformation-review-scientific-literature/ |archive-date=2019-03-06 |url-status=live }}</ref><ref>{{Cite journal|last1=Machado|first1=Caio|last2=Kira|first2=Beatriz|last3=Narayanan|first3=Vidya|last4=Kollanyi|first4=Bence|last5=Howard|first5=Philip|s2cid=153314118|date=2019|title=A Study of Misinformation in WhatsApp groups with a focus on the Brazilian Presidential Elections|journal=Companion Proceedings of the 2019 World Wide Web Conference on – WWW '19|pages=1013–1019|location=New York|publisher=ACM Press|doi=10.1145/3308560.3316738|isbn=978-1450366755}}</ref> Misinformation on social media spreads quickly in comparison to ] because of the lack of regulation and examination required before posting.<ref name=":0" /> These sites provide users with the capability to spread information quickly to other users without requiring the permission of a gatekeeper such as an editor, who might otherwise require confirmation of its truth before allowing its publication. Journalists today are criticized for helping to spread false information on these social platforms, but research such as that from Starbird et al.<ref>{{cite journal |last1=Starbird |first1=Kate |last2=Dailey |first2=Dharma |last3=Mohamed |first3=Owla |last4=Lee |first4=Gina |last5=Spiro |first5=Emma |s2cid=5046314 |title=Engage Early, Correct More: How Journalists Participate in False Rumors Online during Crisis Events |journal=Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (CHI '18) |date=2018 |url=https://www.researchgate.net/publication/322665656|access-date=2019-02-24|doi=10.1145/3173574.3173679 }}</ref> and Arif et al.<ref>{{cite journal |last1=Arif |first1=Ahmer |last2=Robinson |first2=John |last3=Stanck |first3=Stephanie |last4=Fichet |first4=Elodie |last5=Townsend |first5=Paul |last6=Worku |first6=Zena |last7=Starbird |first7=Kate |s2cid=15167363 |title=A Closer Look at the Self-Correcting Crowd: Examining Corrections in Online Rumors |journal=Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing (CSCW '17) |date=2017 |pages=155–169 |doi=10.1145/2998181.2998294 |url=http://ahmerarif.com/papers/CorrectiveBehavior.pdf |accessdate=25 February 2019 |isbn=978-1450343350 |archive-url=https://web.archive.org/web/20190226172742/http://ahmerarif.com/papers/CorrectiveBehavior.pdf |archive-date=26 February 2019 |url-status=live }}</ref> shows they also play a role in curbing the spread of misinformation on social media through debunking and denying false rumors.


== Terminology ==
==Identification and correction==
''Misinformation'' is often used as an umbrella term to refer to many types of false information; more specifically it may refer to false information that is not shared to intentionally deceive or cause harm.<ref>{{Cite web |last=Greason |first=Grace |date=2018-07-06 |title=Information Disorder: The Definitional Toolbox |url=https://firstdraftnews.org/articles/infodisorder-definitional-toolbox/ |access-date=2024-04-03 |website=First Draft }}</ref> Those who do not know that a piece of information is untrue, for instance, might disseminate it on social media in an effort to help.<ref name=":32">{{Cite web |last=Marquez |first=Natalie |title=Research Guides: Misinformation – Get the Facts: What is Misinformation? |url=https://guides.lib.uci.edu/Misinfo/WhatIsMisinfo |access-date=2023-03-16 |website=guides.lib.uci.edu |archive-date=2023-03-16 |archive-url=https://web.archive.org/web/20230316032104/https://guides.lib.uci.edu/Misinfo/WhatIsMisinfo |url-status=live }}</ref>
Information conveyed as credible but later amended can affect people's memory and reasoning after retraction.<ref>{{Cite journal|last1=Ecker|first1=Ullrich K.H.|last2=Lewandowsky|first2=Stephan|last3=Cheung|first3=Candy S.C.|last4=Maybery|first4=Murray T.|date=November 2015|title=He did it! She did it! No, she did not! Multiple causal explanations and the continued influence of misinformation|journal=Journal of Memory and Language|language=en|volume=85|pages=101–115|doi=10.1016/j.jml.2015.09.002|url=https://api.research-repository.uwa.edu.au/files/11787265/Ecker_et_al_He_did_it_She_did_it_2015_.pdf}}</ref> Misinformation differs from concepts like ]s because misinformation is inaccurate information that has previously been disproved.<ref name=":0" /> According to Anne Mintz, editor of ''Web of Deception: Misinformation on the Internet'', the best ways to determine whether information is factual is to use ].<ref>{{cite web | last=Mintz | first=Anne | title=The Misinformation Superhighway? | url=https://www.pbs.org/now/shows/401/internet-politics.html | publisher=PBS | accessdate=26 February 2013 | archive-url=https://web.archive.org/web/20130402040621/http://www.pbs.org/now/shows/401/internet-politics.html | archive-date=2 April 2013 | url-status=live }}</ref> Mintz advises that the reader check whether the information makes sense and whether the founders or reporters of the websites that are spreading the information are biased or have an agenda. Professional journalists and researchers look at other sites (particularly verified sources like news channels<ref>{{Cite book|chapter=Towards automated real-time detection of misinformation on Twitter|publisher= IEEE Conference Publication|pages=2015–2020|language=en-US|doi=10.1109/ICACCI.2016.7732347|date=September 2016|title=2016 International Conference on Advances in Computing, Communications and Informatics (ICACCI)|last1=Jain|first1=Suchita|last2=Sharma|first2=Vanya|last3=Kaushal|first3=Rishabh|s2cid= 17767475|isbn=978-1-5090-2029-4}}</ref>) for that information, as it might be reviewed by multiple people and heavily researched, providing more concrete details.


'']'' is created or spread by a person or organization actively attempting to deceive their audience.<ref name=":42">{{Cite journal |last=Fallis |first=Don |date=2015 |title=What Is Disinformation? |url=http://dx.doi.org/10.1353/lib.2015.0014 |journal=Library Trends |volume=63 |issue=3 |pages=401–426 |doi=10.1353/lib.2015.0014 |hdl=2142/89818 |s2cid=13178809 |issn=1559-0682|hdl-access=free }}</ref> In addition to causing harm directly, disinformation can also cause indirect harm by undermining trust and obstructing the capacity to effectively communicate information with one another.<ref name=":42" /> Disinformation might consist of information that is partially or completely fabricated, taken out of context on purpose, exaggerated, or omits crucial details.<ref name=":52">{{Cite web |title=What is disinformation? |url=https://www.bundesregierung.de/breg-de/themen/umgang-mit-desinformation/disinformation-definition-1911048 |access-date=2023-03-16 |website=Die Bundesregierung informiert {{!}} Startseite |language=de |archive-date=2023-03-23 |archive-url=https://web.archive.org/web/20230323150333/https://www.bundesregierung.de/breg-de/themen/umgang-mit-desinformation/disinformation-definition-1911048 |url-status=live }}</ref> Disinformation can appear in any medium including text, audio, and imagery.<ref name=":52" /> The distinction between mis- and dis-information can be muddy because the intent of someone sharing false information can be difficult to discern.
Martin Libicki, author of ''Conquest In Cyberspace: National Security and Information Warfare'',<ref>{{cite book | last=Libicki | first=Martin | title=Conquest in Cyberspace: National Security and Information Warfare | url=https://archive.org/details/conquestcyberspa00libi_962 | url-access=limited | year=2007 | publisher=Cambridge University Press | location=New York | isbn=978-0521871600 | pages=–55}}</ref> noted that the trick to working with misinformation is the idea that readers must have a balance of what is correct or incorrect. Readers cannot be ] but also should not be ] that all ] is incorrect. There is always a chance that even readers who have this balance will believe an error to be true or that they will disregard factual information as incorrect. According to Libicki, readers' prior ]s or ]s also affect how they interpret new information. When readers believe something to be true before researching it, they are more likely to believe information that supports these prior beliefs or opinions. This phenomenon may lead readers who otherwise are skilled at evaluating credible sources and facts to believe misinformation.


'']'' is accurate information that is disseminated with malicious intent.<ref name=":32" /> This includes sensitive material that is disseminated in order to hurt someone or their reputation.<ref name=":32" /> Examples include ], ], and editing videos to remove important context or content.<ref>{{cite web |title=Foreign Influence Operations and Disinformation {{!}} Cybersecurity and Infrastructure Security Agency CISA |url=https://www.cisa.gov/topics/election-security/foreign-influence-operations-and-disinformation#:~:text=Misinformation%20is%20false%2C%20but%20not,mislead%2C%20harm%2C%20or%20manipulate |website=www.cisa.gov}}</ref>
According to research, the factors that lead to recognizing misinformation is the amount of education a person has and the information literacy they have.<ref>{{Cite journal|last1=Khan|first1=M. Laeeq|last2=Idris|first2=Ika Karlina|s2cid=86681742|date=2019-02-11|title=Recognise misinformation and verify before sharing: a reasoned action and information literacy perspective|journal=Behaviour & Information Technology|volume=38|issue=12|pages=1194–1212|doi=10.1080/0144929x.2019.1578828|issn=0144-929X}}</ref> This means if a person has more knowledge in the subject being investigated, or are familiar with the process of how the information is researched and presented, then they are more likely to identify misinformation. Further research reveal that content descriptors can have a varying effect in people in detecting misinformation.<ref>{{Cite journal|last=Caramancion|first=Kevin Matthe|date=September 2020|title=Understanding the Impact of Contextual Clues in Misinformation Detection|url=https://ieeexplore.ieee.org/document/9216394|journal=2020 IEEE International IOT, Electronics and Mechatronics Conference (IEMTRONICS)|pages=1–6|doi=10.1109/IEMTRONICS51293.2020.9216394|isbn=978-1-7281-9615-2|s2cid=222297695}}</ref>


Misinformation is information that was originally thought to be true but was later discovered not to be true, and often applies to emerging situations in which there is a lack of verifiable information or changing scientific understanding.<ref>{{Cite web |last=Hanopol |first=Nica Rhiana |date=2021-05-12 |title=When 'misinformation' strikes, experts say here's what you can do |url=https://verafiles.org/articles/when-midinformation-strikes-experts-say-heres-what-you-can-d |access-date=2024-04-03 |website=VERA Files }}</ref> For example, the scientific guidance around infant sleep positions has evolved over time,<ref>{{Cite web |title=Safe to Sleep |url=https://safetosleep.nichd.nih.gov/campaign/history |access-date=April 3, 2024}}</ref> and these changes could be a source of confusion for new parents. Misinformation can also often be observed as news events are unfolding and questionable or unverified information fills information gaps. Even if later retracted, false information can continue to influence actions and memory.<ref>{{cite journal |last1=Ecker |first1=Ullrich K.H. |last2=Lewandowsky |first2=Stephan |last3=Cheung |first3=Candy S.C. |last4=Maybery |first4=Murray T. |date=November 2015 |title=He did it! She did it! No, she did not! Multiple causal explanations and the continued influence of misinformation |url=https://api.research-repository.uwa.edu.au/ws/files/11787265/Ecker_et_al_He_did_it_She_did_it_2015_.pdf |url-status=live |journal=Journal of Memory and Language |volume=85 |pages=101–115 |doi=10.1016/j.jml.2015.09.002 |archive-url=https://web.archive.org/web/20221207091937/https://api.research-repository.uwa.edu.au/ws/files/11787265/Ecker_et_al_He_did_it_She_did_it_2015_.pdf |archive-date=2022-12-07 |access-date=2022-11-01}}</ref>
Prior research suggest it can be very difficult to undo the effects of misinformation once individuals believe it to be true and fact checking can even backfire.<ref>{{Cite journal|title=Can Corrections Spread Misinformation to New Audiences? Testing for the Elusive Familiarity Backfire Effect|last1=Ecker|first1=Ullrich K. H.|last2=Lewandowsky|first2=Stephan|date=2020-04-22|last3=Chadwick|first3=Matthew|journal=Cognitive Research: Principles and Implications|volume=5|issue=1|page=41|doi=10.31219/osf.io/et4p3|pmid=32844338|pmc=7447737|url=http://psyarxiv.com/qrm69/}}</ref> Attempting to correct the wrongly held belief is difficult because the misinformation may suit someone's motivational or cognitive reasons. Motivational reasons include the desire to arrive at a foregone conclusion, so accepting information that supports that conclusion. Cognitive reasons may be that the misinformation provides scaffolding for an incident or phenomenon, and is thus part of the mental model for consideration. In this instance, it is necessary to correct the misinformation by not only refuting it, but also by providing accurate information that can also function in the mental model.<ref>{{cite journal|doi=10.1146/annurev-polisci-050718-032814| doi-access=free|title=Political Misinformation|year=2020|last1=Jerit|first1=Jennifer|last2=Zhao|first2=Yangzi|journal=Annual Review of Political Science|volume=23|pages=77–94}}</ref> One suggested solution that would focus on primary prevention of misinformation is the use of a distributed consensus mechanism to validate the accuracy of claims, with appropriate flagging or removal of content that is determined to be false or misleading.<ref>{{cite journal|title=The use of distributed consensus algorithms to curtail the spread of medical misinformation|year=2019|last1=Plaza|first1=Mateusz|last2=Paladino|first2=Lorenzo|s2cid=201803407|journal=International Journal of Academic Medicine|volume=5|issue=2|pages=93–96|doi=10.4103/IJAM.IJAM_47_19}}</ref>


'']s'' are unverified information not attributed to any particular source and may be either true or false.<ref>{{Cite web |date=May 21, 2023 |title=Definition of RUMOR |url=https://www.merriam-webster.com/dictionary/rumor |url-status=live |archive-url=https://web.archive.org/web/20220204010731/https://www.merriam-webster.com/dictionary/rumor |archive-date=February 4, 2022 |access-date=February 4, 2022 |website=www.merriam-webster.com}}</ref>
Flagging or eliminating news media containing false statements using algorithmic fact checkers is becoming the front line in the battle against the spread of misinformation. Computer programs that automatically detect misinformation are still just beginning to emerge, but similar algorithms are already in place with Facebook and Google. Algorithms detect and alert Facebook users that what they are about to share is likely false, hoping to reduce the chances of the user sharing.<ref name=":1">{{Cite journal|last=Lewandowsky|first=Stephan|last2=Ecker|first2=Ullrich K.H.|last3=Cook|first3=John|date=December 2017|title=Beyond Misinformation: Understanding and Coping with the "Post-Truth" Era|url=http://dx.doi.org/10.1016/j.jarmac.2017.07.008|journal=Journal of Applied Research in Memory and Cognition|volume=6|issue=4|pages=353–369|doi=10.1016/j.jarmac.2017.07.008|issn=2211-3681}}</ref> Likewise, Google provides supplemental information pointing to fact check websites in response to its users searching controversial search terms.


Definitions of these terms may vary between cultural contexts.<ref name=":6" />
==Causes==
Historically, people have relied on journalists and other information professionals to relay facts and truths.<ref>{{Cite journal|last=Calvert|first=Philip|date=December 2002|title=Web of Deception: Misinformation on the Internet|journal=The Electronic Library|language=en|volume=20|issue=6|page=521|doi=10.1108/el.2002.20.6.521.7|issn=0264-0473}}</ref> Many different things cause miscommunication but the underlying factor is information literacy. Information is distributed by various means, and because of this it is often hard for users to ask questions of the credibility of what they are seeing. Many online sources of misinformation use techniques to fool users into thinking their sites are legitimate and the information they generate is factual. Often misinformation can be politically motivated. Websites such as USConservativeToday.com have previously posted false information for political and monetary gain.<ref>{{Citation|last=Marwick|first=Alice E.|chapter=Online Identity|date=2013-01-31|work=A Companion to New Media Dynamics|pages=355–364|publisher=Wiley-Blackwell|isbn=978-1-118-32160-7|doi=10.1002/9781118321607.ch23}}</ref> Another role misinformation serves is to distract the public eye from negative information about a given person and/or bigger issues of policy, which as a result can go unremarked with the public preoccupied with fake-news.<ref name=":1" /> In addition to the sharing of misinformation for political and monetary gain it is also spread unintentionally. Advances in digital media have made it easier to share information, although it is not always accurate. The next sections discuss the role social media has in distributing misinformation, the lack of internet gatekeepers, implications of censorship in combating misinformation, inaccurate information from media sources, and competition in news and media.


== History ==
==Social media{{anchor|Social_media_and_misinformation}}==
Early examples include the insults and smears spread among political rivals in Imperial and ] in the form of ]s.<ref>{{Cite web|url=https://www.nybooks.com/daily/2017/02/13/the-true-history-of-fake-news/|title=The True History of Fake News|website=The New York Review of Books|access-date=2019-02-24|date=2017-02-13|archive-url=https://web.archive.org/web/20190205222751/https://www.nybooks.com/daily/2017/02/13/the-true-history-of-fake-news/|archive-date=2019-02-05|url-status=live}}</ref> These are anonymous and witty verses named for the ] piazza and ]. In pre-revolutionary ], "canards", or printed broadsides, sometimes included an engraving to convince readers to take them seriously.{{Citation needed|date=April 2024}}
Contemporary social media platforms offer a rich ground for the spread of misinformation.The exact sharing and motivation behind why misinformation spreads through social media so easily remains unknown.<ref name=":0" /> A 2018 study of Twitter determined that, compared to accurate information, false information spread significantly faster, further, deeper, and more broadly.<ref name="Swire">{{cite journal|doi=10.1146/annurev-publhealth-040119-094127|doi-access=free|title=Public Health and Online Misinformation: Challenges and Recommendations|year=2020|last1=Swire-Thompson|first1=Briony|last2=Lazer|first2=David|journal=Annual Review of Public Health|volume=41|pages=433–451|pmid=31874069}}</ref> Combating its spread is difficult for two reasons: the profusion of information sources, and the generation of "]". The profusion of information sources makes the reader's task of weighing the reliability of information more challenging, heightened by the untrustworthy social signals that go with such information.<ref>{{cite web |last1=Messerole |first1=Chris |title=How misinformation spreads on social media&nbsp;– And what to do about it |url=https://www.brookings.edu/blog/order-from-chaos/2018/05/09/how-misinformation-spreads-on-social-media-and-what-to-do-about-it/ |website=Brookings Institution |accessdate=24 February 2019 |date=2018-05-09 |archive-url=https://web.archive.org/web/20190225044815/https://www.brookings.edu/blog/order-from-chaos/2018/05/09/how-misinformation-spreads-on-social-media-and-what-to-do-about-it/ |archive-date=25 February 2019 |url-status=live }}</ref> The inclination of people to follow or support like-minded individuals leads to the formation of echo chambers and ]s. With no differing information to counter the untruths or the general agreement within isolated social clusters, some writers argue the outcome is a dearth, and worse, the absence of a collective reality, some writers argue.<ref>{{Cite web|url=https://www.cjr.org/analysis/breitbart-media-trump-harvard-study.php|title=Study: Breitbart-led rightwing media ecosystem altered broader media agenda|last=Benkler|first=Y.|date=2017|access-date=8 June 2018|archive-url=https://web.archive.org/web/20180604140114/https://www.cjr.org/analysis/breitbart-media-trump-harvard-study.php|archive-date=4 June 2018|url-status=live}}</ref> Although social media sites have changed their algorithms to prevent the spread of fake news, the problem still exists.<ref>{{Cite web|url=https://web.stanford.edu/~gentzkow/research/fake-news-trends.pdf|title=Trends in the Diffusion of Misinformation on Social Media|last=Allcott|first=Hunt|date=October 2018|website=Stanford Education|bibcode=2018arXiv180905901A|arxiv=1809.05901|access-date=2019-05-10|archive-url=https://web.archive.org/web/20190728160530/https://web.stanford.edu/~gentzkow/research/fake-news-trends.pdf|archive-date=2019-07-28|url-status=live}}</ref> Furthermore, research has shown that while people may know what the scientific community has proved as a fact, they may still refuse to accept it as such.<ref>{{Cite journal|last1=Krause|first1=Nicole M.|last2=Scheufele|first2=Dietram A.|date=2019-04-16|title=Science audiences, misinformation, and fake news|journal=Proceedings of the National Academy of Sciences|language=en|volume=116|issue=16|pages=7662–7669|doi=10.1073/pnas.1805871115|issn=0027-8424|pmid=30642953|pmc=6475373}}</ref>


During the summer of 1587, continental Europe anxiously awaited news as the ] sailed to fight the English. The Spanish postmaster and Spanish agents in Rome promoted reports of Spanish victory in hopes of convincing ] to release his promised one million ]s upon landing of troops. In France, the Spanish and English ambassadors promoted contradictory narratives in the press, and a Spanish victory was incorrectly celebrated in Paris, Prague, and Venice. It was not until late August that reliable reports of the Spanish defeat arrived in major cities and were widely believed; the remains of the fleet returned home in the autumn.<ref>{{cite book |title=The Invention of News |author=Andrew Pettegree |publisher=Yale University Press |year=2015 |isbn=978-0-300-21276-1 |pages=153–4}}</ref>
Misinformation thrives in a social media landscape frequently used and spread by college students.<ref name=":0" /> This can be supported by scholars such as Ghosh and Scott(2018), who indicated that misinformation is "becoming unstoppable".<ref>{{Cite journal|last1=Allcott|first1=Hunt|last2=Gentzkow|first2=Matthew|last3=Yu|first3=Chuan|s2cid=52291737|date=2019-04-01|title=Trends in the diffusion of misinformation on social media|journal=Research & Politics|language=en|volume=6|issue=2|pages=2053168019848554|doi=10.1177/2053168019848554|issn=2053-1680|doi-access=free}}</ref> It has also been observed that misinformation and disinformation come back, multiple times on social media sites. A research study watched the process of thirteen rumors appearing on Twitter and noticed that eleven of those same stories resurfaced multiple times, after much time had passed.<ref>{{Cite journal|last1=Shin|first1=Jieun|last2=Jian|first2=Lian|last3=Driscoll|first3=Kevin|last4=Bar|first4=François|date=June 2018|title=The diffusion of misinformation on social media: Temporal pattern, message, and source|journal=Computers in Human Behavior|volume=83|pages=278–287|doi=10.1016/j.chb.2018.02.008|issn=0747-5632}}</ref>


]
Another reason that misinformation spreads on social media is from the users themselves. In a study, it was shown that the most common reasons that Facebook users were sharing misinformation for social motivated reasons, rather than taking the information seriously.<ref>{{Cite journal|last1=Chen|first1=Xinran|last2=Sin|first2=Sei-Ching Joanna|last3=Theng|first3=Yin-Leng|last4=Lee|first4=Chei Sian|s2cid=15983217|date=2015|title=Why Do Social Media Users Share Misinformation?|journal=Proceedings of the 15th ACM/IEEE-CE on Joint Conference on Digital Libraries – JCDL '15|pages=111–114|location=New York|publisher=ACM Press|doi=10.1145/2756406.2756941|isbn=978-1-4503-3594-2}}</ref> Although users may not be spreading false information for malicious reasons, the misinformation is still being spread across the internet. A research study shows that misinformation that is introduced through a social format influences individuals drastically more than misinformation delivered non-socially.<ref>{{Cite journal|last1=Gabbert|first1=Fiona|last2=Memon|first2=Amina|last3=Allan|first3=Kevin|last4=Wright|first4=Daniel B.|date=September 2004|title=Say it to my face: Examining the effects of socially encountered misinformation|journal=Legal and Criminological Psychology|volume=9|issue=2|pages=215–227|doi=10.1348/1355325041719428|issn=1355-3259|url=https://rke.abertay.ac.uk/ws/files/8509545/Gabbertetal2004Legal.pdf}}</ref>


The first recorded large-scale disinformation campaign was the ], published in 1835 in the New York '']'', in which a series of articles claimed to describe life on the Moon, "complete with illustrations of humanoid bat-creatures and bearded blue unicorns".<ref name="icfj">{{Cite web|title=A short guide to the history of 'fake news' and disinformation|url=https://www.icfj.org/news/short-guide-history-fake-news-and-disinformation-new-icfj-learning-module|url-status=live|archive-url=https://web.archive.org/web/20190225103100/https://www.icfj.org/news/short-guide-history-fake-news-and-disinformation-new-icfj-learning-module|archive-date=2019-02-25|access-date=2019-02-24|website=International Center for Journalists}}</ref> The challenges of mass-producing news on a short deadline can lead to factual errors and mistakes. An example of such is the '']''{{'}}s infamous 1948 headline "]".<ref>{{Cite web |last=Jones |first=Tim |date=2020-10-31 |title=Dewey defeats Truman: The most famous wrong call in electoral history |url=https://www.chicagotribune.com/2020/10/31/dewey-defeats-truman-the-most-famous-wrong-call-in-electoral-history/ |access-date=2024-04-03 |website=Chicago Tribune }}</ref>
Twitter is one of the most concentrated platforms for engagement with political ]. 80% of fake news sources are shared by 0.1% of users, who are "super-sharers". Older, more conservative social users are also more likely to interact with fake news. On Facebook, adults older than 65 were seven times more likely to share fake news than adults ages 18&ndash;29.<ref name="Swire"/>


Social media platforms allow for easy spread of misinformation. Post-election surveys in 2016 suggest that many individuals who intake false information on social media believe them to be factual.<ref>{{Cite journal |last1=Allcott |first1=Hunt |last2=Gentzkow |first2=Matthew |last3=Yu |first3=Chuan |date=April 2019 |title=Trends in the diffusion of misinformation on social media |journal=Research & Politics |volume=6 |issue=2 |pages=205316801984855 |doi=10.1177/2053168019848554 |s2cid=52291737 |issn=2053-1680 |doi-access=free |arxiv=1809.05901 }}</ref> The specific reasons why misinformation spreads through social media so easily remain unknown. A 2018 study of Twitter determined that, compared to accurate information, false information spread significantly faster, further, deeper, and more broadly.<ref name="science.org">{{Cite journal |last1=Vosoughi |first1=Soroush |last2=Roy |first2=Deb |last3=Aral |first3=Sinan |date=2018-03-09 |title=The spread of true and false news online |url=https://www.science.org/doi/10.1126/science.aap9559 |journal=Science |volume=359 |issue=6380 |pages=1146–1151 |doi=10.1126/science.aap9559 |pmid=29590045 |bibcode=2018Sci...359.1146V |issn=0036-8075}}</ref> Similarly, a research study of Facebook found that misinformation was more likely to be clicked on than factual information.{{Citation needed|date=April 2024}}
===Lack of Internet gatekeepers===
] displaying the inaccurate Chicago Tribune headline, an example of misinformation]]
Because of the decentralized nature and structure of the Internet, writers can easily publish content without being required to subject it to ], prove their qualifications, or provide backup documentation. Whereas a book found in a library generally has been reviewed and edited by a second person, Internet sources cannot be assumed to be vetted by anyone other than their authors. They may be produced and posted as soon as the writing is finished.<ref>{{cite journal | last=Stapleton | first=Paul | title=Assessing the quality and bias of web-based sources: implications for academic writing | journal=Journal of English for Academic Purposes | year=2003 | volume=2 | issue=3 | pages=229–245 | doi=10.1016/S1475-1585(03)00026-2}}</ref> In addition, the presence of trolls and bots used to spread willful misinformation has been a problem for social media platforms.<ref>{{Cite news|last=Milman|first=Oliver|url=https://www.theguardian.com/technology/2020/feb/21/climate-tweets-twitter-bots-analysis|title=Revealed: quarter of all tweets about climate crisis produced by bots|date=2020-02-21|work=The Guardian|access-date=2020-02-23|language=en-GB|issn=0261-3077|archive-url=https://web.archive.org/web/20200222193221/https://www.theguardian.com/technology/2020/feb/21/climate-tweets-twitter-bots-analysis|archive-date=2020-02-22|url-status=live}}</ref> As many as 60 million trolls could be actively spreading misinformation on Facebook.<ref>{{Cite journal|last1=Massey|first1=Douglas S.|last2=Iyengar|first2=Shanto|date=2019-04-16|title=Scientific communication in a post-truth society|journal=Proceedings of the National Academy of Sciences|language=en|volume=116|issue=16|pages=7656–7661|doi=10.1073/pnas.1805868115|issn=0027-8424|pmid=30478050|pmc=6475392}}</ref>
Moreover, the advent of the Internet has changed traditional ways that misinformation spreads.<ref>{{Cite journal |last=Godfrey-Smith |first=Peter |date=December 1989 |title=Misinformation |url=https://www.cambridge.org/core/product/identifier/S0045509100016945/type/journal_article |journal=Canadian Journal of Philosophy |volume=19 |issue=4 |pages=533–550 |doi=10.1080/00455091.1989.10716781 |s2cid=246637810 |issn=0045-5091}}</ref> During the ], content from websites deemed 'untrustworthy' reached up to 40% of Americans, despite misinformation making up only 6% of overall news media.<ref>{{Cite journal |last1=West |first1=Jevin D. |last2=Bergstrom |first2=Carl T. |date=2021-04-13 |title=Misinformation in and about science |journal=Proceedings of the National Academy of Sciences |volume=118 |issue=15 |pages=e1912444117 |doi=10.1073/pnas.1912444117 |issn=0027-8424 |pmc=8054004 |pmid=33837146|bibcode=2021PNAS..11812444W |doi-access=free }}</ref> Misinformation has been spread during many health crises.<ref name="NYT-20240316">{{cite news |last=Blum |first=Dani |title=Health Misinformation Is Evolving. Here's How to Spot It. - Experts offer tips for combating false medical claims in your own circles. |url=https://www.nytimes.com/2024/03/16/well/health-misinformation.html |date=March 16, 2024 |work=] |url-status=live |archiveurl=https://archive.today/20240316121000/https://www.nytimes.com/2024/03/16/well/health-misinformation.html |archivedate=March 16, 2024 |accessdate=March 17, 2024 }}</ref><ref name=":6">{{Cite journal |last1=Zeng |first1=Jing |last2=Chan |first2=Chung-hong |date=2021-01-01 |title=A cross-national diagnosis of infodemics: comparing the topical and temporal features of misinformation around COVID-19 in China, India, the US, Germany and France |url=https://doi.org/10.1108/OIR-09-2020-0417 |journal=Online Information Review |volume=45 |issue=4 |pages=709–728 |doi=10.1108/OIR-09-2020-0417 |issn=1468-4527 |s2cid=234242942 |access-date=2023-05-11 |archive-date=2024-02-14 |archive-url=https://web.archive.org/web/20240214210754/https://www.emerald.com/insight/content/doi/10.1108/OIR-09-2020-0417/full/html |url-status=live }}</ref> For example, misinformation about alternative treatments was spread during the Ebola outbreak in 2014–2016.<ref>{{Cite journal |last1=Fung |first1=Isaac Chun-Hai |last2=Fu |first2=King-Wa |last3=Chan |first3=Chung-Hong |last4=Chan |first4=Benedict Shing Bun |last5=Cheung |first5=Chi-Ngai |last6=Abraham |first6=Thomas |last7=Tse |first7=Zion Tsz Ho |date=May 2016 |title=Social Media's Initial Reaction to Information and Misinformation on Ebola, August 2014: Facts and Rumors |journal=Public Health Reports |volume=131 |issue=3 |pages=461–473 |doi=10.1177/003335491613100312 |issn=0033-3549 |pmc=4869079 |pmid=27252566}}</ref><ref>{{Cite journal |last1=Oyeyemi |first1=Sunday Oluwafemi |last2=Gabarron |first2=Elia |last3=Wynn |first3=Rolf |date=2014-10-14 |title=Ebola, Twitter, and misinformation: a dangerous combination? |url=https://www.bmj.com/content/349/bmj.g6178 |journal=BMJ |volume=349 |pages=g6178 |doi=10.1136/bmj.g6178 |issn=1756-1833 |pmid=25315514 |s2cid=731448 |access-date=2023-05-11 |archive-date=2022-11-01 |archive-url=https://web.archive.org/web/20221101202837/https://www.bmj.com/content/349/bmj.g6178 |url-status=live |doi-access=free }}</ref> During the ], the proliferation of mis- and dis-information was exacerbated by a general lack of health literacy.<ref>{{Cite journal |last1=Swire-Thompson |first1=Briony |last2=Lazer |first2=David |date=2020-04-02 |title=Public Health and Online Misinformation: Challenges and Recommendations |journal=Annual Review of Public Health |volume=41 |issue=1 |pages=433–451 |doi=10.1146/annurev-publhealth-040119-094127 |pmid=31874069 |issn=0163-7525|doi-access=free }}</ref>


=== Censorship === == Causes ==
Factors that contribute to beliefs in misinformation are an ongoing subject of study.<ref>{{Cite journal |last1=Jerit |first1=Jennifer |last2=Zhao |first2=Yangzi |date=2020-05-11 |title=Political Misinformation |journal=Annual Review of Political Science |volume=23 |issue=1 |pages=77–94 |doi=10.1146/annurev-polisci-050718-032814 |issn=1094-2939 |s2cid=212733536 |doi-access=free}}</ref> According to ] and Krause, misinformation belief has roots at the individual, group and societal levels.<ref name="Science audiences, misinformation">{{cite journal |last1=Scheufele |first1=Dietram |last2=Krause |first2=Nicole |date=April 16, 2019 |title=Science audiences, misinformation, and fake news |journal=Proceedings of the National Academy of Sciences |volume=116 |issue=16 |pages=7662–7669 |bibcode=2019PNAS..116.7662S |doi=10.1073/pnas.1805871115 |pmc=6475373 |pmid=30642953 |doi-access=free}}</ref> At the individual level, individuals have varying levels of skill in recognizing mis- or dis-information and may be predisposed to certain misinformation beliefs due to other personal beliefs, motivations, or emotions.<ref name="Science audiences, misinformation" /> However, evidence for the hypotheses that believers in misinformation use more cognitive heuristics and less effortful processing of information have produced mixed results.<ref>{{Cite journal |last1=Pennycook |first1=Gordon |last2=Cheyne |first2=James Allan |last3=Seli |first3=Paul |last4=Koehler |first4=Derek J. |last5=Fugelsang |first5=Jonathan A. |date=2012-06-01 |title=Analytic cognitive style predicts religious and paranormal belief |url=https://linkinghub.elsevier.com/retrieve/pii/S0010027712000534 |journal=Cognition |volume=123 |issue=3 |pages=335–346 |doi=10.1016/j.cognition.2012.03.003 |pmid=22481051 |issn=0010-0277}}</ref><ref>{{Cite journal |last1=Robson |first1=Samuel G. |last2=Faasse |first2=Kate |last3=Gordon |first3=Eliza-Rose |last4=Jones |first4=Samuel P. |last5=Drew |first5=Manisara |last6=Martire |first6=Kristy A. |date=2024-10-09 |title=Lazy or different? A quantitative content analysis of how believers and nonbelievers of misinformation reason |url=https://advances.in/psychology/10.56296/aip00027/ |journal=Advances.in/Psychology |language=en |volume=2 |pages=e003511 |doi=10.56296/aip00027 |issn=2976-937X}}</ref><ref>{{Cite journal |last1=Scherer |first1=Laura D. |last2=McPhetres |first2=Jon |last3=Pennycook |first3=Gordon |last4=Kempe |first4=Allison |last5=Allen |first5=Larry A. |last6=Knoepke |first6=Christopher E. |last7=Tate |first7=Channing E. |last8=Matlock |first8=Daniel D. |date=April 2021 |title=Who is susceptible to online health misinformation? A test of four psychosocial hypotheses. |url=http://doi.apa.org/getdoi.cfm?doi=10.1037/hea0000978 |journal=Health Psychology |language=en |volume=40 |issue=4 |pages=274–284 |doi=10.1037/hea0000978 |pmid=33646806 |issn=1930-7810}}</ref> At the group level, ] and a tendency to associate with like-minded or similar people can produce ] and ]s that can create and reinforce misinformation beliefs.<ref name="Science audiences, misinformation" /><ref>{{Cite journal |last=Törnberg |first=Petter |date=2018-09-20 |editor-last=Bauch |editor-first=Chris T. |title=Echo chambers and viral misinformation: Modeling fake news as complex contagion |journal=PLOS ONE |volume=13 |issue=9 |pages=e0203958 |doi=10.1371/journal.pone.0203958 |doi-access=free |issn=1932-6203 |pmc=6147442 |pmid=30235239|bibcode=2018PLoSO..1303958T }}</ref> At the societal level, public figures like politicians and celebrities can disproportionately influence public opinions, as can mass media outlets.<ref>{{cite web |title=Literature Review: Echo chambers, filter bubbles and polarization |url=https://royalsociety.org/-/media/policy/projects/online-information-environment/oie-echo-chambers.pdf |url-status=live |archive-url=https://web.archive.org/web/20220119010245/https://royalsociety.org/-/media/policy/projects/online-information-environment/oie-echo-chambers.pdf |archive-date=19 January 2022 |access-date=21 February 2022}}</ref> In addition, societal trends like political polarization, economic inequalities, declining trust in science, and changing perceptions of authority contribute to the impact of misinformation.<ref name="Science audiences, misinformation" />
Social media sites such as Facebook and Twitter have found themselves defending accusations of censorship for taking down misinformation. In July 2020, a video showing Dr. Stella Immanuel claiming hydroxychloroquine as an effective cure to coronavirus went viral . In the video, Immanuel suggest that there is no need for masks, school closures, or any kind of economic shut down; attesting that this cure she speaks of is highly effective in treating those infected with the virus. The video was shared 600,000 times and received nearly 20 million views on Facebook before it was taken down for violating community guidelines on spreading misinformation.<ref name=":2">{{Cite news|date=2020-07-29|title=Stella Immanuel - the doctor behind unproven coronavirus cure claim|language=en-GB|work=BBC News|url=https://www.bbc.com/news/world-africa-53579773|access-date=2020-11-23}}</ref> The video was also taken down on Twitter overnight, but not before President Donald Trump shared it to his page, which is followed by over 85 million Twitter users.<ref name=":2" /> NIAID director Dr. Anthony Fauci and members of the World Health Organization (WHO) quickly discredited the video citing larger scale studies of hydroxychloroquine showing it is not an effective treatment of Covid-19, and the FDA cautioned against using it to treat Covid patients following evidence of serious heart problems arising in patients that have taken the drug.<ref name=":2" /> Facebook and Twitter alike have policies in place to combat misinformation regarding Covid-19, and the social media platforms were swift with action to back up their respective policies.


Historically, people have relied on journalists and other information professionals to relay facts.<ref>{{cite journal |title=Web of Deception: Misinformation on the Internet |journal=The Electronic Library |date=1 December 2002 |volume=20 |issue=6 |pages=521 |doi=10.1108/el.2002.20.6.521.7 }}</ref> As the number and variety of information sources has increased, it has become more challenging for the general public to assess their ].<ref>{{Cite web |title=Students Have 'Dismaying' Inability To Tell Fake News From Real, Study Finds |website=] |url=https://www.npr.org/sections/thetwo-way/2016/11/23/503129818/study-finds-students-have-dismaying-inability-to-tell-fake-news-from-real |access-date=April 4, 2024}}</ref> This growth of consumer choice when it comes to news media allows the consumer to choose a news source that may align with their biases, which consequently increases the likelihood that they are misinformed.<ref name="BeyondMisinfo" /> 47% of Americans reported social media as their main news source in 2017 as opposed to traditional news sources.<ref>{{Cite web |last1=Shearer |first1=Elisa |last2=Gottfried |first2=Jeffrey |date=2017-09-07 |title=News Use Across Social Media Platforms 2017 |url=https://www.journalism.org/2017/09/07/news-use-across-social-media-platforms-2017/ |url-status=live |archive-url=https://web.archive.org/web/20210316054814/https://www.journalism.org/2017/09/07/news-use-across-social-media-platforms-2017/ |archive-date=2021-03-16 |access-date=2021-03-28 |website=Pew Research Center's Journalism Project}}</ref> Polling shows that Americans trust mass media at record-low rates,<ref>{{Cite web |date=2023-10-19 |title=Media Confidence in U.S. Matches 2016 Record Low |url=https://news.gallup.com/poll/512861/media-confidence-matches-2016-record-low.aspx |access-date=2024-05-16 |website=Gallup.com |language=en}}</ref> and that US young adults place similar levels of trust in information from social media and from national news organizations.<ref>{{Cite web |last1=Liedke |first1=Jacob |last2=Gottfried |first2=Jeffrey |date=2022-10-27 |title=U.S. adults under 30 now trust information from social media almost as much as from national news outlets |url=https://www.pewresearch.org/short-reads/2022/10/27/u-s-adults-under-30-now-trust-information-from-social-media-almost-as-much-as-from-national-news-outlets/ |access-date=2024-05-16 |website=Pew Research Center |language=en-US}}</ref> The pace of the ] does not always allow for adequate ], potentially leading to the spread of misinformation.<ref>{{cite book |last1=Croteau |first1=David |title=Media Society: Industries, Images, and Audiences |last2=Hoynes |first2=William |last3=Milan |first3=Stefania |pages=285–321 |chapter=Media Technology |access-date=March 21, 2013 |chapter-url=http://www.sagepub.com/upm-data/40857_9.pdf |archive-url=https://web.archive.org/web/20130102172415/http://www.sagepub.com/upm-data/40857_9.pdf |archive-date=January 2, 2013 |url-status=live}}</ref> Further, the distinction between opinion and reporting can be unclear to viewers or readers.<ref>{{Cite web |last1=Mitchell |first1=Amy |last2=Gottfried |first2=Jeffrey |last3=Barthel |first3=Michael |last4=Sumida |first4=Nami |date=2018-06-18 |title=Distinguishing Between Factual and Opinion Statements in the News |url=https://www.pewresearch.org/journalism/2018/06/18/distinguishing-between-factual-and-opinion-statements-in-the-news/ |access-date=2024-05-16 |website=Pew Research Center |language=en-US}}</ref><ref>{{Cite web |title=Journalists know news and opinion are separate, but readers often can't tell the difference |url=https://www.niemanlab.org/2020/06/journalists-know-news-and-opinion-are-separate-but-readers-often-cant-tell-the-difference/ |access-date=2024-05-16 |website=Nieman Lab}}</ref>
===Inaccurate information from media sources===
A Gallup poll made public in 2016 found that only 32% of Americans trust the mass media "to report the news fully, accurately and fairly", the lowest number in the history of that poll.<ref name=":3">{{Cite book|last=Marwick|first=Alice|title=Media Manipulation and Disinformation Online|last2=Lewis|first2=Rebecca|publisher=Data & Society Research Institute|year=2017|isbn=|location=New York|pages=40–45}}</ref> An example of bad information from media sources that led to the spread of misinformation occurred in November 2005, when Chris Hansen on '']'' made a claim that ] officials estimate 50,000 predators are online at any moment. Afterwards, the U.S. attorney general at the time, Alberto Gonzales, repeated the claim. However, the number that Hansen used in his reporting had no backing. Hansen said he received the information from ''Dateline'' expert Ken Lanning, but Lanning admitted that he made up the number 50,000 because there was no solid data on the number. According to Lanning, he used 50,000 because it sounds like a real number, not too big and not too small, and referred to it as a "] number". Reporter Carl Bialik says that the number 50,000 is used often in the media to estimate numbers when reporters are unsure of the exact data.<ref>{{Cite book|title=The Influencing Machine|last=Gladstone|first=Brooke|publisher=W. W. Norton & Company|year=2012|isbn=978-0393342468|location=New York|pages=49–51}}</ref>


Sources of misinformation can appear highly convincing and similar to trusted legitimate sources.<ref>{{Cite web |publisher=Stanford University |date=2016-11-21 |title=Stanford researchers find students have trouble judging the credibility of information online |url=https://ed.stanford.edu/news/stanford-researchers-find-students-have-trouble-judging-credibility-information-online |access-date=2024-04-03 |website=Stanford Graduate School of Education }}</ref> For example, misinformation cited with hyperlinks has been found to increase readers' trust. Trust is even higher when these hyperlinks are to scientific journals, and higher still when readers do not click on the sources to investigate for themselves.<ref>{{cite journal |last1=Verma |first1=Nitin |last2=Fleischmann |first2=Kenneth R. |last3=Koltai |first3=Kolina S. |date=January 2017 |title=Human values and trust in scientific journals, the mainstream media and fake news |journal=Proceedings of the Association for Information Science and Technology |volume=54 |issue=1 |pages=426–435 |doi=10.1002/pra2.2017.14505401046 |s2cid=51958978}}</ref><ref>{{Citation |last1=Sehgal |first1=Vibhor |title=Mutual Hyperlinking Among Misinformation Peddlers |date=2021-04-20 |arxiv=2104.11694 |last2=Peshin |first2=Ankit |last3=Afroz |first3=Sadia |last4=Farid |first4=Hany}}</ref> Research has also shown that the presence of relevant images alongside incorrect statements increases both their believability and shareability, even if the images do not actually provide evidence for the statements.<ref name=":14">{{Cite journal |last1=Newman |first1=Eryn J. |last2=Garry |first2=Maryanne |last3=Bernstein |first3=Daniel M. |last4=Kantner |first4=Justin |last5=Lindsay |first5=D. Stephen |date=October 2012 |title=Nonprobative photographs (or words) inflate truthiness |url=https://pubmed.ncbi.nlm.nih.gov/22869334 |journal=Psychonomic Bulletin & Review |volume=19 |issue=5 |pages=969–974 |doi=10.3758/s13423-012-0292-0 |issn=1531-5320 |pmid=22869334}}</ref><ref>{{Cite journal |last1=Sundar |first1=S Shyam |last2=Molina |first2=Maria D |last3=Cho |first3=Eugene |date=2021-11-12 |title=Seeing Is Believing: Is Video Modality More Powerful in Spreading Fake News via Online Messaging Apps? |url=https://academic.oup.com/jcmc/article/26/6/301/6336055 |journal=Journal of Computer-Mediated Communication |volume=26 |issue=6 |pages=301–319 |doi=10.1093/jcmc/zmab010 |issn=1083-6101|doi-access=free }}</ref> For example, a false statement about macadamia nuts accompanied by an image of a bowl of macadamia nuts tends to be rated as more believable than the same statement without an image.<ref name=":14" />
===Competition in news and media===
Because news organizations and websites hotly compete for viewers, there is a need for great efficiency in releasing stories to the public. The news media landscape in the 1970's offered American consumers access to a limited, but overall consistent and trusted selection of news offerings, where as today consumers are confronted with an overabundance of voices online.<ref name=":1" /> This explosion of consumer choice when it comes to news media allows the consumer to pick and choose a news source that hits their preferred agenda, which consequently increases the likelihood that they are misinformed.<ref name=":1" /> ] companies broadcast stories 24 hours a day, and break the latest news in hopes of taking audience share from their competitors. News is also produced at a pace that does not always allow for ], or for all of the facts to be collected or released to the media at one time, letting readers or viewers insert their own opinions, and possibly leading to the spread of misinformation.<ref>{{cite journal | last=Croteau | display-authors=etal | title=Media Technology | pages=285–321 | url=http://www.sagepub.com/upm-data/40857_9.pdf | accessdate=March 21, 2013 | journal= | archive-url=https://web.archive.org/web/20130102172415/http://www.sagepub.com/upm-data/40857_9.pdf | archive-date=January 2, 2013 | url-status=live }}</ref>


The translation of scientific research into popular reporting can also lead to confusion if it flattens nuance, sensationalizes the findings, or places too much emphasis on weaker levels of ]. For instance, researchers have found that newspapers are more likely than scientific journals to cover observational studies and studies with weaker methodologies.<ref>{{Cite journal |last1=Selvaraj |first1=Senthil |last2=Borkar |first2=Durga S. |last3=Prasad |first3=Vinay |date=2014-01-17 |title=Media Coverage of Medical Journals: Do the Best Articles Make the News? |journal=PLOS ONE |volume=9 |issue=1 |pages=e85355 |doi=10.1371/journal.pone.0085355 |doi-access=free |issn=1932-6203 |pmc=3894978 |pmid=24465543|bibcode=2014PLoSO...985355S }}</ref> Dramatic headlines may gain readers' attention, but they do not always accurately reflect scientific findings.<ref>{{Cite web |title=How to spot misleading science reporting – QB3 Berkeley |url=https://qb3.berkeley.edu/news/how-to-spot-misleading-science-reporting/ |access-date=2024-04-12 |website=qb3.berkeley.edu}}</ref>
==Impact==
Misinformation can affect all aspects of life. Allcott, Gentzkow and Yu (2019:6) concur that diffusion of misinformation through social media is a potential threat to democracy and broader society. The effects of misinformation can lead to the accurateness about information and details of the occurrence to decline.<ref>{{Cite journal|last1=Bodner|first1=Glen E.|last2=Musch|first2=Elisabeth|last3=Azad|first3=Tanjeem|date=2009|title=Reevaluating the potency of the memory conformity effect|journal=Memory & Cognition|volume=37|issue=8|pages=1069–1076|doi=10.3758/mc.37.8.1069|issn=0090-502X|pmid=19933452|doi-access=free}}</ref> When ] on conversations, one can gather facts that may not always be true, or the receiver may hear the message incorrectly and spread the information to others. On the Internet, one can read content that is stated to be factual but that may not have been checked or may be erroneous. In the news, companies may emphasize the speed at which they receive and send information but may not always be correct in the facts. These developments contribute to the way misinformation will continue to complicate the public's understanding of issues and to serve as a source for belief and attitude formation.<ref>{{Cite book|url=https://books.google.com/books?id=1Jo7DwAAQBAJ&q=misinformation+causes&pg=PT246|title=Misinformation and Mass Audiences|last1=Southwell|first1=Brian G.|last2=Thorson|first2=Emily A.|last3=Sheble|first3=Laura|date=2018|publisher=University of Texas Press|isbn=978-1477314586|language=en}}</ref>


Human cognitive tendencies can also be a contributing factor to misinformation belief. One study found that an individual's recollection of political events could be altered when presented with misinformation about the event, even when primed to identify warning signs of misinformation.<ref>{{Cite journal |last1=Freeze |first1=Melanie |last2=Baumgartner |first2=Mary |last3=Bruno |first3=Peter |last4=Gunderson |first4=Jacob R. |last5=Olin |first5=Joshua |last6=Ross |first6=Morgan Quinn |last7=Szafran |first7=Justine |date=2021-12-01 |title=Fake Claims of Fake News: Political Misinformation, Warnings, and the Tainted Truth Effect |journal=Political Behavior |volume=43 |issue=4 |pages=1433–1465 |doi=10.1007/s11109-020-09597-3 |issn=1573-6687|doi-access=free }}</ref> Misinformation may also be appealing by seeming ] or incorporating existing ]s.<ref name="Healy 2020">{{cite news |last1=Healy |first1=Melissa |date=8 February 2020 |title=Misinformation about the coronavirus abounds, but correcting it can backfire |url=https://www.latimes.com/science/story/2020-02-08/coronavirus-outbreak-false-information-psychology |work=Los Angeles Times}}</ref>
In regards to ], some view being a misinformed citizen as worse than being an uninformed citizen. Misinformed citizens can state their beliefs and opinions with confidence and in turn affect elections and policies. This type of misinformation comes from speakers not always being upfront and straightforward, yet may appear both "authoritative and legitimate" on the surface.<ref name="The Growing Role of Social Media in"/> When information is presented as vague, ambiguous, sarcastic, or partial, receivers are forced to piece the information together and make assumptions about what is correct.<ref>{{cite book | last=Barker | first=David | title=Rushed to Judgement: Talk Radio, Persuasion, and American Political Behavior | year=2002 | publisher=Columbia University Press | location=New York | pages=106–109}}</ref> Aside from ], misinformation can also be employed in industrial propaganda. Using tools such as advertising, a company can undermine reliable evidence or influence belief through concerted misinformation campaign. For instance, ] companies employed misinformation in the second half of the twentieth century to diminish the reliability of studies that demonstrated the link between ] and ].<ref>{{Cite book|title=The Misinformation Age: How False Beliefs Spread|url=https://archive.org/details/misinformationag0000ocon|url-access=limited|last1=O'Connor|first1=Cailin|last2=Weatherall|first2=James Owen|publisher=Yale University Press|year=2019|isbn=978-0300234015|location=New Haven|pages=}}</ref> In the medical field, misinformation can immediately lead to life endangerment as seen in the case of the public's negative perception towards vaccines or the use of herbs instead of medicines to treat diseases.<ref name="The Growing Role of Social Media in"/><ref>{{Cite book|url=https://books.google.com/books?id=9ZuMDwAAQBAJ&q=misinformation+causes&pg=PT22|title=India Misinformed: The True Story|last1=Sinha|first1=P.|last2=Shaikh|first2=S.|last3=Sidharth|first3=A.|publisher=Harper Collins|year=2019|isbn=978-9353028381|location=|pages=}}</ref> In regards to the COVID-19 pandemic, the spread of misinformation has proven to cause confusion as well as negative emotions such as anxiety and fear.<ref name="Bratu 2020">{{cite journal |last1=Bratu |first1=Sofia |title=The Fake News Sociology of COVID-19 Pandemic Fear: Dangerously Inaccurate Beliefs, Emotional Contagion, and Conspiracy Ideation |journal=Linguistic and Philosophical Investigations |date=May 24, 2020 |volume=19 |pages=128–134 |doi=10.22381/LPI19202010 |url=https://doi-org.ez.salem.edu/10.22381/LPI19202010 |accessdate=7 November 2020}}</ref> Misinformation regarding proper safety measures for the prevention of the virus that go against information from legitimate institutions like the World Health Organization can also lead to inadequate protection and possibly place individuals at risk for exposure.<ref name="Bratu 2020"/><ref>{{Cite web|date=2020-07-29|title=Misinformation on coronavirus is proving highly contagious|url=https://apnews.com/article/ap-top-news-understanding-the-outbreak-health-media-social-media-86f61f3ffb6173c29bc7db201c10f141|access-date=2020-11-23|website=AP NEWS}}</ref>


== Identification ==
Misinformation has the power to sway public elections and referendums if it has the chance to gain enough momentum in the public discourse. Leading up to the ] for example, a figure widely circulated by the ] campaign claimed the UK would save £350 million a week by leaving the EU, and that the money would be redistributed to the British ].<ref name=":4">{{Cite web|date=2018-07-27|title=The misinformation that was told about Brexit during and after the referendum|url=https://www.independent.co.uk/news/uk/politics/final-say-brexit-referendum-lies-boris-johnson-leave-campaign-remain-a8466751.html|access-date=2020-11-23|website=The Independent|language=en}}</ref> This was later deemed a "clear misuse of official statistics" by the UK statistics authority.<ref name=":4" /> The advert infamously shown off on the side of London's renowned double decker busses did not take into account the UK's budget rebate, and the idea that 100% of the money saved would go to the NHS was unrealistic. A poll published in 2016 by ] found that nearly half of the British public believed this misinformation to be true.<ref name=":4" /> Even when information is proven to me misinformation, it may continue to shape people's attitudes towards a given topic,<ref name=":3" /> meaning misinformation has the power to swing political decisions if it gains enough traction in public discussion.
Research has yielded a number of strategies that can be employed to identify misinformation, many of which share common features. According to Anne Mintz, editor of ''Web of Deception: Misinformation on the Internet'', one of the simplest ways to determine whether information is factual is to use ].<ref>{{cite web | last=Mintz | first=Anne | title=The Misinformation Superhighway? | url=https://www.pbs.org/now/shows/401/internet-politics.html | publisher=PBS | access-date=26 February 2013 | archive-url=https://web.archive.org/web/20130402040621/http://www.pbs.org/now/shows/401/internet-politics.html | archive-date=2 April 2013 | url-status=live }}</ref> Mintz advises that the reader check whether the information makes sense and whether the source or sharers of the information might be ] or have an agenda. However, because emotions and preconceptions heavily impact belief, this is not always a reliable strategy.<ref name=":16">{{Cite journal |last1=Ecker |first1=Ullrich K. H. |last2=Lewandowsky |first2=Stephan |last3=Cook |first3=John |last4=Schmid |first4=Philipp |last5=Fazio |first5=Lisa K. |last6=Brashier |first6=Nadia |last7=Kendeou |first7=Panayiota |last8=Vraga |first8=Emily K. |last9=Amazeen |first9=Michelle A. |date=January 2022 |title=The psychological drivers of misinformation belief and its resistance to correction |url=https://www.nature.com/articles/s44159-021-00006-y |journal=Nature Reviews Psychology |volume=1 |issue=1 |pages=13–29 |doi=10.1038/s44159-021-00006-y |issn=2731-0574|hdl=1983/889ddb0f-0d44-44f4-a54f-57c260ae4917 |hdl-access=free }}</ref> Readers tend to distinguish between unintentional misinformation and uncertain evidence from politically or financially motivated misinformation.<ref name="g426"/> The perception of misinformation depends on the ], with right-wing readers more concerned with attempts to hide reality.<ref name="g426">{{cite journal | last1=Hameleers | first1=Michael | last2=Brosius | first2=Anna | title=You Are Wrong Because I Am Right! The Perceived Causes and Ideological Biases of Misinformation Beliefs | journal=International Journal of Public Opinion Research | volume=34 | issue=1 | date=16 February 2022 | issn=0954-2892 | doi=10.1093/ijpor/edab028 | page=}}</ref> It can be difficult to undo the effects of misinformation once individuals believe it to be true.<ref name="CognitiveResearch">{{Cite journal |last1=Ecker |first1=Ullrich K. H. |last2=Lewandowsky |first2=Stephan |last3=Chadwick |first3=Matthew |date=2020-04-22 |title=Can Corrections Spread Misinformation to New Audiences? Testing for the Elusive Familiarity Backfire Effect |url=http://psyarxiv.com/qrm69/ |url-status=live |journal=Cognitive Research: Principles and Implications |volume=5 |issue=1 |page=41 |doi=10.31219/osf.io/et4p3 |pmc=7447737 |pmid=32844338 |archive-url=https://web.archive.org/web/20200617162551/https://psyarxiv.com/qrm69/ |archive-date=2020-06-17 |access-date=2020-06-17 |hdl=1983/0d5feec2-5878-4af6-b5c7-fbbd398dd4c4}}</ref> Individuals may desire to reach a certain conclusion, causing them to accept information that supports that conclusion, and are more likely to retain and share information if it emotionally resonates with them.<ref>{{Cite journal |last1=Lewandowsky |first1=Stephan |last2=Ecker |first2=Ullrich K. H. |last3=Seifert |first3=Colleen M. |last4=Schwarz |first4=Norbert |last5=Cook |first5=John |year=2012 |title=Misinformation and Its Correction: Continued Influence and Successful Debiasing |journal=Psychological Science in the Public Interest |volume=13 |issue=3 |pages=106–131 |doi=10.1177/1529100612451018 |jstor=23484653 |pmid=26173286 |s2cid=42633 |doi-access=free}}</ref>


The SIFT Method, also called the Four Moves, is one commonly taught method of distinguishing between reliable and unreliable information.<ref>{{Cite web |date=2019-06-19 |title=SIFT (The Four Moves) |url=https://hapgood.us/2019/06/19/sift-the-four-moves/ |access-date=2024-04-12 |website=Hapgood }}</ref> This method instructs readers to first Stop and begin to ask themselves about what they are reading or viewing - do they know the source and if it is reliable? Second, readers should Investigate the source. What is the source's relevant expertise and do they have an agenda? Third, a reader should Find better coverage and look for reliable coverage on the claim at hand to understand if there is a consensus around the issue. Finally, a reader should Trace claims, quotes, or media to their original context: has important information been omitted, or is the original source questionable?
Websites have been created to help people to discern fact from fiction. For example, the site ] has a mission to fact check the media, especially politician speeches and stories going viral on the Internet. The site also includes a forum where people can openly ask questions about information they're not sure is true in both the media and the internet.<ref>{{Cite web|url=http://www.factcheck.org/askfactcheck/|title=Ask FactCheck|website=www.factcheck.org|access-date=2016-03-31|archive-url=https://web.archive.org/web/20160331063044/http://www.factcheck.org/askfactcheck|archive-date=2016-03-31|url-status=live}}</ref> Similar sites give individuals the option to be able to copy and paste misinformation into a search engine and the site will investigate the truthfulness of the inputted data.<ref name="Fernandez 2018">{{Cite journal|last1=Fernandez|first1=Miriam|last2=Alani|first2=Harith|s2cid=13799324|date=2018|title=Online Misinformation|journal=Companion of the Web Conference 2018 on the Web Conference 2018 – WWW '18|pages=595–602|location=New York|publisher=ACM Press|doi=10.1145/3184558.3188730|isbn=978-1-4503-5640-4|url=http://oro.open.ac.uk/53734/1/sample-sigconf.pdf|access-date=2020-02-13|archive-url=https://web.archive.org/web/20190411195506/http://oro.open.ac.uk/53734/1/sample-sigconf.pdf|archive-date=2019-04-11|url-status=live}}</ref> Famous online resources, such as Facebook and Google, have attempted to add automatic fact checker programs to their sites, and created the option for users to flag information that they think are false on their website.<ref name="Fernandez 2018"/> A way that fact checking programs find misinformation involve finding the truth by analyzing the language and syntax of news stories. Another way is that fact checkers can search for existing information on the subject and compare it to the new broadcasts being put online.<ref>{{Cite journal|last1=Zhang|first1=Chaowei|last2=Gupta|first2=Ashish|last3=Kauten|first3=Christian|last4=Deokar|first4=Amit V.|last5=Qin|first5=Xiao|date=December 2019|title=Detecting fake news for reducing misinformation risks using analytics approaches|journal=European Journal of Operational Research|volume=279|issue=3|pages=1036–1052|doi=10.1016/j.ejor.2019.06.022|issn=0377-2217}}</ref> Other sites such as Misplaced Pages and ] are also widely used resources for verifying information.


Visual misinformation presents particular challenges, but there are some effective strategies for identification.<ref>{{Cite journal |last1=Cowles |first1=Kelsey |last2=Miller |first2=Rebekah |last3=Suppok |first3=Rachel |date=2024-01-02 |title=When Seeing Isn't Believing: Navigating Visual Health Misinformation through Library Instruction |url=https://www.tandfonline.com/doi/full/10.1080/02763869.2024.2290963 |journal=Medical Reference Services Quarterly |volume=43 |issue=1 |pages=44–58 |doi=10.1080/02763869.2024.2290963 |pmid=38237023 |issn=0276-3869}}</ref> Misleading graphs and charts can be identified through careful examination of the data presentation; for example, truncated axes or poor color choices can cause confusion.<ref>{{Cite web |last=Grootendorst |first=Maarten |date=2021-03-23 |title=Misleading Graphs |url=https://www.maartengrootendorst.com/blog/misleading/ |access-date=2024-04-12 |website=Maarten Grootendorst }}</ref> Reverse image searching can reveal whether images have been taken out of their original context.<ref>{{Cite journal |last1=Qian |first1=Sijia |last2=Shen |first2=Cuihua |last3=Zhang |first3=Jingwen |date=2022-11-04 |title=Fighting cheapfakes: using a digital media literacy intervention to motivate reverse search of out-of-context visual misinformation |journal=Journal of Computer-Mediated Communication |volume=28 |issue=1 |doi=10.1093/jcmc/zmac024 |issn=1083-6101|doi-access=free }}</ref> There are currently some somewhat reliable ways to identify ]-generated imagery,<ref>{{Cite journal |last1=Groh |first1=Matthew |last2=Epstein |first2=Ziv |last3=Firestone |first3=Chaz |last4=Picard |first4=Rosalind |date=2022-01-05 |title=Deepfake detection by human crowds, machines, and machine-informed crowds |journal=Proceedings of the National Academy of Sciences |volume=119 |issue=1 |doi=10.1073/pnas.2110013119 |doi-access=free |issn=0027-8424 |pmc=8740705 |pmid=34969837|arxiv=2105.06496 |bibcode=2022PNAS..11910013G }}</ref><ref>{{Cite web |last=Bond |first=Shannon |title=AI-generated images are everywhere. Here's how to spot them |website=] |url=https://www.npr.org/2023/06/07/1180768459/how-to-identify-ai-generated-deepfake-images |access-date=April 12, 2024}}</ref> but it is likely that this will become more difficult to identify as the technology advances.<ref>{{Cite news |last=Hern |first=Alex |date=2024-04-08 |title='Time is running out': can a future of undetectable deepfakes be avoided? |url=https://www.theguardian.com/technology/2024/apr/08/time-is-running-out-can-a-future-of-undetectable-deepfakes-be-avoided |access-date=2024-04-12 |work=The Guardian |issn=0261-3077}}</ref><ref>{{Cite news |last=Verma |first=Pranshu |title=AI Can Draw Hands Now |newspaper=] |url=https://www.washingtonpost.com/technology/2023/03/26/ai-generated-hands-midjourney/}}</ref>
Some scholars and activists are pioneering a movement to eliminate the mis/disinformation and information pollution in the digital world. The theory they are developing, "information environmentalism", has become a curriculum in some universities and colleges.<ref>{{Cite news|url=https://er.educause.edu/articles/2017/10/info-environmentalism-an-introduction|title=Info-Environmentalism: An Introduction|access-date=2018-09-28|language=en|archive-url=https://web.archive.org/web/20180703130624/https://er.educause.edu/articles/2017/10/info-environmentalism-an-introduction|archive-date=2018-07-03|url-status=live}}</ref><ref>{{Cite news|url=https://dlinq.middcreate.net/informationenvironmentalism/|title=Information Environmentalism|date=2017-12-21|work=Digital Learning and Inquiry (DLINQ)|access-date=2018-09-28|language=en|archive-url=https://web.archive.org/web/20180928044347/https://dlinq.middcreate.net/informationenvironmentalism/|archive-date=2018-09-28|url-status=live}}</ref>


A person's formal education level and ] do correlate with their ability to recognize misinformation.<ref>{{cite journal |last1=Khan |first1=M. Laeeq |last2=Idris |first2=Ika Karlina |date=2 December 2019 |title=Recognise misinformation and verify before sharing: a reasoned action and information literacy perspective |journal=Behaviour & Information Technology |volume=38 |issue=12 |pages=1194–1212 |doi=10.1080/0144929x.2019.1578828 |s2cid=86681742}}</ref><ref name="sfn">{{Cite journal |last1=Lazer |first1=David M. J. |last2=Baum |first2=Matthew A. |last3=Benkler |first3=Yochai |last4=Berinsky |first4=Adam J. |last5=Greenhill |first5=Kelly M. |last6=Menczer |first6=Filippo |last7=Metzger |first7=Miriam J. |last8=Nyhan |first8=Brendan |last9=Pennycook |first9=Gordon |last10=Rothschild |first10=David |last11=Schudson |first11=Michael |last12=Sloman |first12=Steven A. |last13=Sunstein |first13=Cass R. |last14=Thorson |first14=Emily A. |last15=Watts |first15=Duncan J. |year=2018 |title=The science of fake news |journal=Science |volume=359 |issue=6380 |pages=1094–1096 |arxiv=2307.07903 |bibcode=2018Sci...359.1094L |doi=10.1126/science.aao2998 |pmid=29590025 |s2cid=4410672 |last16=Zittrain |first16=Jonathan L.}}</ref> People who are familiar with a topic, the processes of researching and presenting information, or have critical evaluation skills are more likely to correctly identify misinformation. However, these are not always direct relationships. Higher overall literacy does not always lead to improved ability to detect misinformation.<ref>{{cite journal |last1=Vraga |first1=Emily K. |last2=Bode |first2=Leticia |date=December 2017 |title=Leveraging institutions, educators, and networks to correct misinformation: A commentary on Lewandosky, Ecker, and Cook. |journal=Journal of Applied Research in Memory and Cognition |volume=6 |issue=4 |pages=382–388 |doi=10.1016/j.jarmac.2017.09.008}}</ref> Context clues can also significantly impact people's ability to detect misinformation.<ref>{{cite book |last1=Caramancion |first1=Kevin Matthe |title=2020 IEEE International IOT, Electronics and Mechatronics Conference (IEMTRONICS) |year=2020 |isbn=978-1-7281-9615-2 |pages=1–6 |chapter=Understanding the Impact of Contextual Clues in Misinformation Detection |doi=10.1109/IEMTRONICS51293.2020.9216394 |s2cid=222297695}}</ref>
==See also==

{{columns-list|colwidth=30em|
], author of ''Conquest In Cyberspace: National Security and Information Warfare'',<ref>{{cite book | last=Libicki | first=Martin | title=Conquest in Cyberspace: National Security and Information Warfare | url=https://archive.org/details/conquestcyberspa00libi_962 | url-access=limited | year=2007 | publisher=Cambridge University Press | location=New York | isbn=978-0-521-87160-0 | pages=–55}}</ref> notes that readers should aim to be skeptical but not cynical. Readers should not be ], believing everything they read without question, but also should not be ] that everything they see or read is false.

==Countermeasures{{anchor|Countering misinformation}}==
{{see also|Conspiracy theories#Interventions}}
Factors that contribute to the effectiveness of a corrective message include an individual's mental model or ], repeated exposure to the misinformation, time between misinformation and correction, credibility of the sources, and relative coherency of the misinformation and corrective message. Corrective messages will be more effective when they are coherent and/or consistent with the audience's worldview. They will be less effective when misinformation is believed to come from a credible source, is repeated prior to correction (even if the repetition occurs in the process of debunking), and/or when there is a time lag between the misinformation exposure and corrective message. Additionally, corrective messages delivered by the original source of the misinformation tend to be more effective.<ref>{{cite journal |last1=Walter |first1=Nathan |last2=Tukachinsky |first2=Riva |title=A Meta-Analytic Examination of the Continued Influence of Misinformation in the Face of Correction: How Powerful Is It, Why Does It Happen, and How to Stop It? |journal=Communication Research |date=March 2020 |volume=47 |issue=2 |pages=155–177 |doi=10.1177/0093650219854600 |s2cid=197731687 |url=https://digitalcommons.chapman.edu/cgi/viewcontent.cgi?article=1061&context=comm_articles |access-date=2022-11-01 |archive-date=2022-12-07 |archive-url=https://web.archive.org/web/20221207092441/https://digitalcommons.chapman.edu/cgi/viewcontent.cgi?article=1061&context=comm_articles |url-status=live }}</ref> However, misinformation research has often been criticized for its emphasis on efficacy (i.e., demonstrating effects of interventions in controlled experiments) over effectiveness (i.e., confirming real-world impacts of these interventions).<ref name=":13">{{Cite journal |last1=Roozenbeek |first1=Jon |last2=Remshard |first2=Miriam |last3=Kyrychenko |first3=Yara |date=2024-07-27 |title=Beyond the headlines: On the efficacy and effectiveness of misinformation interventions |url=https://advances.in/psychology/10.56296/aip00019/ |journal=Advances.in/Psychology |language=en |volume=2 |pages=e24569 |doi=10.56296/aip00019 |issn=2976-937X}}</ref> Critics argue that while laboratory settings may show promising results, these do not always translate into practical, everyday situations where misinformation spreads.<ref>{{Cite journal |last1=Tay |first1=Li Qian |last2=Lewandowsky |first2=Stephan |last3=Hurlstone |first3=Mark J. |last4=Kurz |first4=Tim |last5=Ecker |first5=Ullrich K. H. |date=2023-10-05 |title=A focus shift in the evaluation of misinformation interventions |url=https://misinforeview.hks.harvard.edu/article/a-focus-shift-in-the-evaluation-of-misinformation-interventions/ |journal=Harvard Kennedy School Misinformation Review |volume=4 |issue=5 |language=en-US |doi=10.37016/mr-2020-124|pmid=39301480 |pmc=7616579 }}</ref> Research has identified several major challenges in this field: an overabundance of lab research and a lack of field studies, the presence of testing effects that impede intervention longevity and scalability, modest effects for small fractions of relevant audiences, reliance on item evaluation tasks as primary efficacy measures, low replicability in the Global South and a lack of audience-tailored interventions, and the underappreciation of potential unintended consequences of intervention implementation.<ref name=":13" />

=== Fact-checking and debunking ===
Websites have been created to help people to discern fact from fiction. For example, the site ] aims to fact check the media, especially ] political stories. The site also includes a forum where people can openly ask questions about the information.<ref>{{Cite web |title=Ask FactCheck |url=http://www.factcheck.org/askfactcheck/ |url-status=live |archive-url=https://web.archive.org/web/20160331063044/http://www.factcheck.org/askfactcheck |archive-date=2016-03-31 |access-date=2016-03-31 |website=www.factcheck.org}}</ref> Similar sites allow individuals to copy and paste misinformation into a search engine and the site will investigate it.<ref name=":12">{{cite book |last1=Fernandez |first1=Miriam |title=Companion of the Web Conference 2018 on the Web Conference 2018 – WWW '18 |last2=Alani |first2=Harith |publisher=ACM Press |year=2018 |isbn=978-1-4503-5640-4 |pages=595–602 |chapter=Online Misinformation |doi=10.1145/3184558.3188730 |s2cid=13799324}}</ref> Some sites exist to address misinformation about specific topics, such as climate change misinformation. ], formerly The DeSmogBlog, publishes factually accurate information in order to counter the well-funded disinformation campaigns spread by motivated ]. ] focuses on evaluating science, health, climate, and energy claims in the media and providing an evidence-based analysis of their veracity.<ref>{{Cite web |title=About - Science Feedback |url=https://science.feedback.org/about/ |access-date=2024-04-12 |website=science.feedback.org/ }}</ref>

Flagging or eliminating false statements in media using algorithmic fact checkers is becoming an increasingly common tactic to fight misinformation. Google and many social media platforms have added automatic fact-checking programs to their sites and created the option for users to flag information that they think is false.<ref name=":12" /> Google provides supplemental information pointing to fact-checking websites in search results for controversial topics. On Facebook, algorithms may warn users if what they are about to share is likely false.<ref name="BeyondMisinfo">{{cite journal |last1=Lewandowsky |first1=Stephan |last2=Ecker |first2=Ullrich K. H. |last3=Cook |first3=John |title=Beyond misinformation: Understanding and coping with the 'post-truth' era |journal=Journal of Applied Research in Memory and Cognition |date=December 2017 |volume=6 |issue=4 |pages=353–369 |doi=10.1016/j.jarmac.2017.07.008 |hdl=1983/1b4da4f3-009d-4287-8e45-a0a1d7b688f7 |s2cid=149003083 |url=https://api.research-repository.uwa.edu.au/ws/files/22250770/Lewandowsky.2017.JARMAC.pdf |access-date=2022-11-01 |archive-date=2023-01-18 |archive-url=https://web.archive.org/web/20230118190604/https://api.research-repository.uwa.edu.au/ws/files/22250770/Lewandowsky.2017.JARMAC.pdf |url-status=live }}</ref> In some cases social media platforms' efforts to curb the spread of misinformation has resulted in controversy, drawing criticism from people who see these efforts as constructing a barrier to their right to expression.<ref>{{Cite web |last=Nott |first=Lata |date=2022-08-03 |title=Free Speech on Social Media: The Complete Guide |url=https://www.freedomforum.org/free-speech-on-social-media/ |access-date=2024-04-12 |website=Freedom Forum }}</ref>

==== One-on-one correction ====
Within the context of personal interactions, some strategies for debunking have the potential to be effective. Simply delivering facts is frequently ineffective because misinformation belief is often not the result of a deficit of accurate information,<ref name=":16" /> although individuals may be more likely to change their beliefs in response to information shared by someone with whom they have close social ties, like a friend or family member.<ref>{{Cite journal |last1=Pasquetto |first1=Irene V. |last2=Jahani |first2=Eaman |last3=Atreja |first3=Shubham |last4=Baum |first4=Matthew |date=2022-04-07 |title=Social Debunking of Misinformation on WhatsApp: The Case for Strong and In-group Ties |url=https://doi.org/10.1145/3512964 |journal=Proceedings of the ACM on Human-Computer Interaction |volume=6 |issue=CSCW1 |pages=117:1–117:35 |doi=10.1145/3512964}}</ref> More effective strategies focus on instilling doubt and encouraging people to examine the roots of their beliefs.<ref>{{Cite news |date=2020-12-21 |title=How should you talk to friends and relatives who believe conspiracy theories? |url=https://www.bbc.com/news/blogs-trending-55350794 |access-date=2024-04-22 }}</ref> In these situations, tone can also play a role: expressing empathy and understanding can keep communication channels open. It is important to remember that beliefs are driven not just by facts but by emotion, worldview, intuition, ], and many other factors.<ref name=":16" />

==== Social correction ====
Fact-checking and debunking can be done in one-on-one interactions, but when this occurs on social media it is likely that other people may encounter and read the interaction, potentially learning new information from it or examining their own beliefs. This type of correction has been termed social correction.<ref name=":18">{{Cite journal |last1=Vraga |first1=Emily K. |last2=Bode |first2=Leticia |date=October 2020 |title=Correction as a Solution for Health Misinformation on Social Media |journal=American Journal of Public Health |volume=110 |issue=S3 |pages=S278–S280 |doi=10.2105/AJPH.2020.305916 |issn=0090-0036 |pmc=7532323 |pmid=33001724}}</ref> Researchers have identified three ways to increase the efficacy of these social corrections for observers.<ref name=":18" /> First, corrections should include a link to a credible source of relevant information, like an expert organization. Second, the correct information should be repeated, for example at the beginning and end of the comment or response. Third, an alternative explanation should be offered. An effective social correction in response to a statement that chili peppers can cure COVID-19 might look something like: “Hot peppers in your food, though very tasty, cannot prevent or cure COVID-19. The best way to protect yourself against the new coronavirus is to keep at least 1 meter away from others and to wash your hands frequently and thoroughly. Adding peppers to your soup won't prevent or cure COVID-19. Learn more from the WHO."<ref>{{Cite web |title=COVID-19 Mythbusters – World Health Organization |url=https://www.who.int/emergencies/diseases/novel-coronavirus-2019/advice-for-public/myth-busters |access-date=2024-04-22 |website=www.who.int }}</ref> Interestingly, while the tone of the correction may impact how the target of the correction receives the message and can increase engagement with a message,<ref>{{Cite journal |last1=MacKay |first1=Melissa |last2=Cimino |first2=Andrea |last3=Yousefinaghani |first3=Samira |last4=McWhirter |first4=Jennifer E. |last5=Dara |first5=Rozita |last6=Papadopoulos |first6=Andrew |date=2022-06-06 |title=Canadian COVID-19 Crisis Communication on Twitter: Mixed Methods Research Examining Tweets from Government, Politicians, and Public Health for Crisis Communication Guiding Principles and Tweet Engagement |journal=International Journal of Environmental Research and Public Health |volume=19 |issue=11 |pages=6954 |doi=10.3390/ijerph19116954 |doi-access=free |issn=1660-4601 |pmc=9180105 |pmid=35682537}}</ref> it is less likely to affect how others seeing the correction perceive its accuracy.<ref>{{Cite journal |last1=Bode |first1=Leticia |last2=Vraga |first2=Emily K. |last3=Tully |first3=Melissa |date=2020-06-11 |title=Do the right thing: Tone may not affect correction of misinformation on social media |url=https://misinforeview.hks.harvard.edu/article/do-the-right-thing-tone-may-not-affect-correction-of-misinformation-on-social-media/ |journal=Harvard Kennedy School Misinformation Review |doi=10.37016/mr-2020-026|doi-access=free }}</ref>

While social correction has the potential to reach a wider audience with correct information, it can also potentially amplify an original post containing misinformation.<ref>{{Cite web |title=Health Misinformation — Current Priorities of the U.S. Surgeon General |url=https://www.hhs.gov/surgeongeneral/priorities/health-misinformation/index.html |access-date=2024-04-22 |website=www.hhs.gov }}</ref>

=== Prebunking ===
Unfortunately, misinformation typically spreads more readily than fact-checking.<ref name=":11" /><ref>{{Cite web |last=Smith |first=Rory |date=2020-12-18 |title=The "broadcast" model no longer works in an era of disinformation |url=https://firstdraftnews.org/articles/the-broadcast-model-no-longer-works-in-an-era-of-disinformation/ |access-date=2024-04-12 |website=First Draft }}</ref><ref name="science.org"/> Further, even if misinformation is corrected, that does not mean it is forgotten or does not influence people's thoughts.<ref name=":11" /> Another approach, called prebunking, aims to "inoculate" against misinformation by showing people examples of misinformation and how it works before they encounter it.<ref name=":17">{{Cite web |last=Garcia |first=Laura |date=2021-06-29 |title=A guide to prebunking: a promising way to inoculate against misinformation |url=https://firstdraftnews.org/articles/a-guide-to-prebunking-a-promising-way-to-inoculate-against-misinformation/ |access-date=2024-04-12 |website=First Draft }}</ref><ref>{{cite journal |last1=Cook |first1=John |last2=Ellerton |first2=Peter |last3=Kinkead |first3=David |date=1 February 2018 |title=Deconstructing climate misinformation to identify reasoning errors |journal=Environmental Research Letters |volume=13 |issue=2 |pages=024018 |bibcode=2018ERL....13b4018C |doi=10.1088/1748-9326/aaa49f |s2cid=149353744 |doi-access=free}}</ref> While prebunking can involve fact-based correction, it focuses more on identifying common logical fallacies (e.g., emotional appeals to manipulate individuals' perceptions and judgments,<ref>{{Cite journal |last1=Traberg |first1=Cecilie |last2=Morton |first2=Thomas |last3=van der Linden |first3=Sander |date=2024-06-20 |title=Counteracting socially endorsed misinformation through an emotion-fallacy inoculation |url=https://advances.in/psychology/10.56296/aip00017/ |journal=Advances.in/Psychology |language=en |volume=2 |pages=e765332 |doi=10.56296/aip00017 |issn=2976-937X}}</ref> ], or ] fallacies<ref>{{Cite journal |last1=Roozenbeek |first1=Jon |last2=van der Linden |first2=Sander |last3=Goldberg |first3=Beth |last4=Rathje |first4=Steve |last5=Lewandowsky |first5=Stephan |date=2022-08-26 |title=Psychological inoculation improves resilience against misinformation on social media |journal=Science Advances |language=en |volume=8 |issue=34 |pages=eabo6254 |doi=10.1126/sciadv.abo6254 |issn=2375-2548 |pmc=9401631 |pmid=36001675|bibcode=2022SciA....8O6254R }}</ref>) and tactics used to spread misinformation as well as common misinformation sources.<ref name=":17" /> Research about the efficacy of prebunking has shown promising results.<ref>{{Cite journal |last1=Roozenbeek |first1=Jon |last2=van der Linden |first2=Sander |date=2019-06-25 |title=Fake news game confers psychological resistance against online misinformation |journal=Palgrave Communications |volume=5 |issue=1 |pages=1–10 |doi=10.1057/s41599-019-0279-9 |issn=2055-1045|doi-access=free }}</ref>

=== Other interventions ===
A report by the ] in the UK lists additional potential or proposed countermeasures:<ref name="onlenv" />
* Automated detection systems (e.g. to flag or add context and resources to content)
* Provenance enhancing technology (i.e. better enabling people to determine the veracity of a claim, image, or video)
* APIs for research (i.e. for usage to detect, understand, and counter misinformation)
* Active bystanders (e.g. corrective commenting)
* Community moderation (usually of unpaid and untrained, often independent, volunteers)
* Anti-virals (e.g. limiting the number of times a message can be forwarded in privacy-respecting encrypted chats)
* ] (examples being Misplaced Pages where multiple editors refine encyclopedic articles, and question-and-answer sites where outputs are also evaluated by others similar to ])
* Trustworthy institutions and data
* ] (increasing citizens' ability to use ] to find, evaluate, create, and communicate information, an essential skill for citizens of all ages)
** Media literacy is taught in ]n public schools – from kindergarten through to high school – since 2010 and "accepted 'as important as writing or reading'"<ref>{{cite news |last1=Yee |first1=Amy |title=The country inoculating against disinformation |url=https://www.bbc.com/future/article/20220128-the-country-inoculating-against-disinformation |access-date=21 February 2022 |work=BBC |archive-date=20 February 2022 |archive-url=https://web.archive.org/web/20220220115030/https://www.bbc.com/future/article/20220128-the-country-inoculating-against-disinformation |url-status=live }}</ref>
** ] mandated K-12 students to learn information literacy<ref>{{cite web |last1=Sitrin |first1=Carly |title=New Jersey becomes first state to mandate K-12 students learn information literacy |url=https://www.politico.com/news/2023/01/05/new-jersey-is-the-first-state-to-mandate-k-12-students-learn-information-literacy-00076352 |website=Politico |date=5 January 2023 |access-date=9 January 2023 |archive-date=9 January 2023 |archive-url=https://web.archive.org/web/20230109111755/https://www.politico.com/news/2023/01/05/new-jersey-is-the-first-state-to-mandate-k-12-students-learn-information-literacy-00076352 |url-status=live }}</ref>
** "Inoculation" via educational videos shown to adults is being explored<ref>{{cite journal |last1=Roozenbeek |first1=Jon |last2=van der Linden |first2=Sander |last3=Goldberg |first3=Beth |last4=Rathje |first4=Steve |last5=Lewandowsky |first5=Stephan |title=Psychological inoculation improves resilience against misinformation on social media |journal=Science Advances |date=26 August 2022 |volume=8 |issue=34 |pages=eabo6254 |doi=10.1126/sciadv.abo6254 |pmid=36001675 |pmc=9401631 |bibcode=2022SciA....8O6254R |issn=2375-2548}}</ref>

Broadly described, the report recommends building resilience to scientific misinformation and a healthy online information environment and not having offending content removed. It cautions that censorship could e.g. drive misinformation and associated communities "to harder-to-address corners of the internet".<ref>{{cite web|title=Royal Society cautions against censorship of scientific misinformation online|work=]|url=https://royalsociety.org/news/2022/01/scientific-misinformation-report/|access-date=12 February 2022|archive-date=12 February 2022|archive-url=https://web.archive.org/web/20220212113919/https://royalsociety.org/news/2022/01/scientific-misinformation-report/|url-status=live}}</ref>

Online ] can be counteracted through different measures at different stages.<ref>{{cite journal |last1=Treen |first1=Kathie M. d'I. |last2=Williams |first2=Hywel T. P. |last3=O'Neill |first3=Saffron J. |title=Online misinformation about climate change |journal=WIREs Climate Change |date=September 2020 |volume=11 |issue=5 |doi=10.1002/wcc.665 |s2cid=221879878 |doi-access=free |bibcode=2020WIRCC..11E.665T }}</ref> Prior to misinformation exposure, education and "inoculation" are proposed. Technological solutions, such as early detection of bots and ranking and selection algorithms are suggested as ongoing mechanisms. Post misinformation, corrective and collaborator messaging can be used to counter climate change misinformation. Incorporating fines and similar consequences has also been suggested.

The ] was launched in 2023 as a consortium of over 250 scientists working to develop effective countermeasures to misinformation and other problems created by perverse incentives in organizations disseminating information via the Internet.<ref><!-- National Academies of Sciences, Engineering, and Medicine (2023) Nobel Prize Summit Fuels Initiatives to Combat Misinformation and Disinformation and Build Trust in Science-->{{cite Q|Q124711722|author=National Academies of Sciences, Engineering, and Medicine}}</ref>

There also is research and development of platform-built-in as well ]-integrated (currently in the form of ]) misinformation mitigation.<ref>{{cite news |last1=Zewe |first1=Adam |title=Empowering social media users to assess content helps fight misinformation |url=https://techxplore.com/news/2022-11-empowering-social-media-users-content.html |access-date=18 December 2022 |work=Massachusetts Institute of Technology via techxplore.com |archive-date=18 December 2022 |archive-url=https://web.archive.org/web/20221218171535/https://techxplore.com/news/2022-11-empowering-social-media-users-content.html |url-status=live }}</ref><ref>{{cite journal |last1=Jahanbakhsh |first1=Farnaz |last2=Zhang |first2=Amy X. |last3=Karger |first3=David R. |title=Leveraging Structured Trusted-Peer Assessments to Combat Misinformation |journal=Proceedings of the ACM on Human-Computer Interaction |date=11 November 2022 |volume=6 |issue=CSCW2 |pages=524:1–524:40 |doi=10.1145/3555637|doi-access=free|hdl=1721.1/147638 |hdl-access=free }}</ref><ref>{{cite web |last1=Elliott |first1=Matt |title=Fake news spotter: How to enable Microsoft Edge's NewsGuard |url=https://www.cnet.com/tech/mobile/fake-news-spotter-how-to-enable-microsoft-edges-newsguard/ |website=CNET |access-date=9 January 2023 |archive-date=9 January 2023 |archive-url=https://web.archive.org/web/20230109225348/https://www.cnet.com/tech/mobile/fake-news-spotter-how-to-enable-microsoft-edges-newsguard/ |url-status=live }}</ref><ref>{{cite web |title=12 Browser Extensions to Help You Detect and Avoid Fake News |url=https://thetrustedweb.org/browser-extensions-to-detect-and-avoid-fake-news/ |website=The Trusted Web |access-date=9 January 2023 |date=18 March 2021 |archive-date=9 January 2023 |archive-url=https://web.archive.org/web/20230109225350/https://thetrustedweb.org/browser-extensions-to-detect-and-avoid-fake-news/ |url-status=live }}</ref> This includes quality/neutrality/reliability ratings for news sources. Misplaced Pages's perennial sources page categorizes many large news sources by reliability.<ref>{{cite news |last1=Darcy |first1=Oliver |title=Misplaced Pages administrators caution editors about using Fox News as source on 'contentious' claims {{!}} CNN Business |url=https://edition.cnn.com/2020/07/24/media/wikipedia-fox-news-reliable-sources/index.html |access-date=9 January 2023 |work=CNN |date=24 July 2020 |archive-date=20 November 2021 |archive-url=https://web.archive.org/web/20211120115036/https://edition.cnn.com/2020/07/24/media/wikipedia-fox-news-reliable-sources/index.html |url-status=live }}</ref> Researchers have also demonstrated the feasibility of ] for popular and official figures by developing such for over 800 contemporary ]s on Twitter as well as associated exposure scores.<ref>{{cite news |title=New MIT Sloan research measures exposure to misinformation from political elites on Twitter |url=https://apnews.com/press-release/pr-newswire/misinformation-701fb46656eb2197a845f789857d83b2 |access-date=18 December 2022 |work=AP NEWS |date=29 November 2022 |archive-date=18 December 2022 |archive-url=https://web.archive.org/web/20221218171535/https://apnews.com/press-release/pr-newswire/misinformation-701fb46656eb2197a845f789857d83b2 |url-status=live }}</ref><ref>{{cite journal |last1=Mosleh |first1=Mohsen |last2=Rand |first2=David G. |title=Measuring exposure to misinformation from political elites on Twitter |journal=Nature Communications |date=21 November 2022 |volume=13 |issue=1 |pages=7144 |doi=10.1038/s41467-022-34769-6 |pmid=36414634 |pmc=9681735 |bibcode=2022NatCo..13.7144M |issn=2041-1723|doi-access=free}}</ref>

Strategies that may be more effective for lasting correction of false beliefs include focusing on intermediaries (such as convincing activists or politicians who are credible to the people who hold false beliefs, or promoting intermediaries who have the same identities or worldviews as the intended audience), minimizing the association of misinformation with political or group identities (such as providing corrections from nonpartisan experts, or avoiding ] based on partisanship in news coverage), and emphasizing corrections that are hard for people to avoid or deny (such as providing information that the economy is unusually strong or weak, or describing the increased occurrence of ] events in response to ]).<ref name="Nyhan2021">{{cite journal |author=Nyhan B |year=2021 |title=Why the backfire effect does not explain the durability of political misperceptions. |journal=Proc Natl Acad Sci U S A |volume=118 |issue=15 |pages= |bibcode=2021PNAS..11812440N |doi=10.1073/pnas.1912440117 |pmc=8053951 |pmid=33837144 |doi-access=free}}</ref>

===Limitations===
{{main|Belief perseverance}}

Interventions need to account for the possibility that misinformation can persist in the population even after corrections are published. Possible reasons include difficulty in reaching the right people and corrections not having long-term effects.<ref name="Nyhan2021"/><ref name=":13"/> For example, if corrective information is only published in science-focused publications and fact-checking websites, it may not reach the people who believe in misinformation since they are less likely to read those sources. In addition, successful corrections may not be persistent, particularly if people are re-exposed to misinformation at a later date.<ref name="Nyhan2021" />

It has been suggested that directly countering misinformation can be counterproductive, which is referred to as a "backfire effect", but in practice this is very rare.<ref name="Nyhan2021"/><ref name="LazicZezelj2021">{{cite journal | last1=Lazić | first1=Aleksandra | last2=Žeželj | first2=Iris | title=A systematic review of narrative interventions: Lessons for countering anti-vaccination conspiracy theories and misinformation | journal=Public Understanding of Science | publisher=SAGE Publications | volume=30 | issue=6 | date=18 May 2021 | issn=0963-6625 | doi=10.1177/09636625211011881 | pages=644–670| pmid=34006153 }}</ref><ref name="Swire-Thompson2020"/><ref>{{Citation|last1=Lewandowsky|first1=Stephan|title=Debunking Handbook 2020|date=2020|url=http://databrary.org/volume/1182|publisher=Databrary|doi=10.17910/b7.1182|access-date=2021-01-20|last2=Cook|first2=John|last3=Lombardi|first3=Doug|pages=9–11}}</ref> A 2020 review of the scientific literature on backfire effects found that there have been widespread ] their existence, even under conditions that would be theoretically favorable to observing them.<ref name="Swire-Thompson2020">{{cite journal| vauthors=Swire-Thompson B, DeGutis J, Lazer D| title=Searching for the Backfire Effect: Measurement and Design Considerations. | journal=J Appl Res Mem Cogn | year= 2020 | volume= 9 | issue= 3 | pages= 286–299 | pmid=32905023 | doi=10.1016/j.jarmac.2020.06.006 | pmc=7462781 }}</ref> Due to the lack of ], {{as of|2020|lc=y}} most researchers believe that backfire effects are either unlikely to occur on the broader population level, or they only occur in very specific circumstances, or they do not exist.<ref name="Swire-Thompson2020"/> Brendan Nyhan, one of the researchers who initially proposed the occurrence of backfire effects, wrote in 2021 that the persistence of misinformation is most likely due to other factors.<ref name="Nyhan2021"/> For most people, corrections and fact-checking are very unlikely to have a negative impact, and there is no specific group of people in which backfire effects have been consistently observed.<ref name="Swire-Thompson2020"/> In many cases, when backfire effects have been discussed by the media or by bloggers, they have been ] from studies on specific subgroups to incorrectly conclude that backfire effects apply to the entire population and to all attempts at correction.<ref name="Nyhan2021"/><ref name="Swire-Thompson2020"/>

There is an ongoing debate on whether misinformation interventions may have the negative side effect of reducing belief in both false and true information, regardless of veracity.<ref>{{Cite journal |last1=Guay |first1=Brian |last2=Berinsky |first2=Adam J. |last3=Pennycook |first3=Gordon |last4=Rand |first4=David |date=August 2023 |title=How to think about whether misinformation interventions work |url=https://www.nature.com/articles/s41562-023-01667-w |journal=Nature Human Behaviour |language=en |volume=7 |issue=8 |pages=1231–1233 |doi=10.1038/s41562-023-01667-w |pmid=37563304 |issn=2397-3374}}</ref> For instance, one study found that inoculation and accuracy primes to some extent undermined users' ability to distinguish implausible from plausible conspiracy theories.<ref>{{Cite journal |last1=O'Mahony |first1=Cian |last2=Murphy |first2=Gillian |last3=Linehan |first3=Conor |date=2024-11-15 |title=True discernment or blind scepticism? Comparing the effectiveness of four conspiracy belief interventions |url=https://advances.in/psychology/10.56296/aip00030/ |journal=Advances.in/Psychology |language=en |volume=2 |pages=e215691 |doi=10.56296/aip00030 |issn=2976-937X}}</ref> Other scholars have shown through simulations that even if interventions reduce both the belief in false and true information, the effect on the media ecosystem may still be favorable due to different base rates in both beliefs.<ref>{{Cite journal |last1=Lawson |first1=M. Asher |last2=Kakkar |first2=Hemant |date=2024-12-06 |title=Resolving conflicting findings in misinformation research: A methodological perspective |url=https://advances.in/psychology/10.56296/aip00031/ |journal=Advances.in/Psychology |language=en |volume=2 |pages=e235462 |doi=10.56296/aip00031 |issn=2976-937X}}</ref>

== Online misinformation{{anchor|Social_media_and_misinformation}} ==
{{further|Health information on the Internet#Social media}}
In recent years, the proliferation of misinformation online has drawn widespread attention.<ref name=":02">{{Cite journal |last=Acerbi |first=Alberto |date=2019-02-12 |title=Cognitive attraction and online misinformation |journal=Palgrave Communications |volume=5 |issue=1 |doi=10.1057/s41599-019-0224-y |s2cid=257089832 |issn=2055-1045|doi-access=free |hdl=11572/357769 |hdl-access=free }}</ref> More than half of the world's population had access to the Internet in the beginning of 2018.<ref name=":02" /> Digital and social media can contribute to the spread of misinformation – for instance, when users share information without first checking the legitimacy of the information they have found. People are more likely to encounter online information based on personalized algorithms.<ref name=":12"/> Google, Facebook and Yahoo News all generate newsfeeds based on the information they know about our devices, our location, and our online interests.<ref name=":12" />

Although two people can search for the same thing at the same time, they are very likely to get different results based on what that platform deems relevant to their interests, fact or false.<ref name=":12" /> Various social media platforms have recently been criticized for encouraging the spread of false information, such as hoaxes, false news, and mistruths.<ref name=":12" /> It is responsible with influencing people's attitudes and judgment during significant events by disseminating widely believed misinformation.<ref name=":12" /> Furthermore, online misinformation can occur in numerous ways, including rumors, urban legends, factoids, etc.<ref name=":22">{{Cite book |last=Almaliki |first=Malik |date=2019-04-06 |chapter=Online Misinformation Spread |title=Proceedings of the 2019 3rd International Conference on Information System and Data Mining |pages=171–178 |location=New York |publisher=ACM |doi=10.1145/3325917.3325938|isbn=978-1-4503-6635-9 |s2cid=190232270 }}</ref> However, the underlying factor is that it contains misleading or inaccurate information.<ref name=":22" />

Moreover, users of social media platforms may experience intensely negative feelings, perplexity, and worry as a result of the spread of false information.<ref name=":22" /> According to a recent study, one in ten Americans has gone through mental or emotional stress as a result of misleading information posted online.<ref name=":22" /> Spreading false information can also seriously impede the effective and efficient use of the information available on social media.<ref name=":22" /> An emerging trend in the online information environment is "a shift away from public discourse to private, more ephemeral, ]", which is a challenge to counter misinformation.<ref name="onlenv">{{cite book |title=The online information environment: Understanding how the internet shapes people's engagement with scientific information |date=January 2022 |isbn=978-1-78252-567-7 |publisher=The Royal Society |url=https://royalsociety.org/-/media/policy/projects/online-information-environment/the-online-information-environment.pdf |access-date=21 February 2022 |archive-date=11 February 2022 |archive-url=https://web.archive.org/web/20220211131118/https://royalsociety.org/-/media/policy/projects/online-information-environment/the-online-information-environment.pdf |url-status=live }}</ref>

=== On social media ===
Pew Research reports shared that approximately one in four American adults admitted to sharing misinformation on their social media platforms.<ref>{{Cite web |last1=Barthel |first1=Michael |last2=Mitchell |first2=Amy |last3=Holcomb |first3=Jesse |date=2016-12-15 |title=Many Americans Believe Fake News Is Sowing Confusion |url=https://www.pewresearch.org/journalism/2016/12/15/many-americans-believe-fake-news-is-sowing-confusion/ |access-date=2024-04-03 |website=Pew Research Center's Journalism Project }}</ref>

In the ], ] have become a notable agent for the spread of misinformation, ], and propaganda.<ref name="The Growing Role of Social Media in">{{cite book |doi=10.1007/978-3-030-23491-1_14 |chapter=The Growing Role of Social Media in International Health Security: The Good, the Bad, and the Ugly |title=Global Health Security |series=Advanced Sciences and Technologies for Security Applications |year=2020 |last1=Stawicki |first1=Stanislaw P. |last2=Firstenberg |first2=Michael S. |last3=Papadimos |first3=Thomas J. |pages=341–357 |isbn=978-3-030-23490-4 |s2cid=212995901 }}</ref><ref name="sfn" /><ref name="vosoughi">{{cite journal |last1=Vosoughi |first1=Soroush |last2=Roy |first2=Deb |last3=Aral |first3=Sinan |title=The spread of true and false news online |journal=Science |date=9 March 2018 |volume=359 |issue=6380 |pages=1146–1151 |doi=10.1126/science.aap9559 |pmid=29590045 |bibcode=2018Sci...359.1146V |s2cid=4549072 |url=http://vermontcomplexsystems.org/share/papershredder/vosoughi2018a.pdf |access-date=21 August 2019 |archive-date=29 April 2019 |archive-url=https://web.archive.org/web/20190429073158/http://vermontcomplexsystems.org/share/papershredder/vosoughi2018a.pdf |url-status=dead }}</ref><ref name="tucker">{{cite journal|last1=Tucker|first1=Joshua A.|last2=Guess|first2=Andrew|last3=Barbera|first3=Pablo|last4=Vaccari|first4=Cristian|last5=Siegel|first5=Alexandra|last6=Sanovich|first6=Sergey|last7=Stukal|first7=Denis|last8=Nyhan|first8=Brendan|title=Social Media, Political Polarization, and Political Disinformation: A Review of the Scientific Literature|url=https://hewlett.org/library/social-media-political-polarization-political-disinformation-review-scientific-literature/|url-status=live|journal=Hewlett Foundation White Paper|archive-url=https://web.archive.org/web/20190306111605/https://hewlett.org/library/social-media-political-polarization-political-disinformation-review-scientific-literature/|archive-date=2019-03-06|access-date=2019-03-05}}</ref><ref>{{cite book |doi=10.1145/3308560.3316738 |chapter=A Study of Misinformation in WhatsApp groups with a focus on the Brazilian Presidential Elections |title=Companion Proceedings of the 2019 World Wide Web Conference |year=2019 |last1=Machado |first1=Caio |last2=Kira |first2=Beatriz |last3=Narayanan |first3=Vidya |last4=Kollanyi |first4=Bence |last5=Howard |first5=Philip |pages=1013–1019 |isbn=978-1-4503-6675-5 |s2cid=153314118 }}</ref> Social media sites have changed their algorithms to prevent the spread of ] but the problem still exists.<ref name="Allcott Gentzkow Yu 2019" />

Image posts are the biggest spread of misinformation on social media, a fact which is grossly unrepresented in research. This leads to a "yawning gap of knowledge" as there is a collective ignorance on how harmful image-based posts are compared to other types of misinformation.<ref>{{Cite web |last1=Hindman |first1=Matthew |last2=Davis |first2=Trevor |last3=Yang |first3=Yunkang |date=2023-06-30 |title=Visual misinformation is widespread on Facebook – and often undercounted by researchers |url=http://theconversation.com/visual-misinformation-is-widespread-on-facebook-and-often-undercounted-by-researchers-202913 |access-date=2024-01-25 |website=The Conversation |archive-date=2024-01-25 |archive-url=https://web.archive.org/web/20240125142416/http://theconversation.com/visual-misinformation-is-widespread-on-facebook-and-often-undercounted-by-researchers-202913 |url-status=live }}</ref>

==== Spread ====
Social media platforms allow for easy spread of misinformation.<ref name="Allcott Gentzkow Yu 2019">{{cite journal |last1=Allcott |first1=Hunt |last2=Gentzkow |first2=Matthew |last3=Yu |first3=Chuan |title=Trends in the diffusion of misinformation on social media |journal=Research & Politics |date=April 2019 |volume=6 |issue=2 |pages=205316801984855 |doi=10.1177/2053168019848554 |s2cid=52291737 |doi-access=free |arxiv=1809.05901 }}</ref> The specific reasons why misinformation spreads through social media so easily remain unknown.<ref name="TJAL">{{cite journal |last1=Chen |first1=Xinran |last2=Sin |first2=Sei-Ching Joanna |last3=Theng |first3=Yin-Leng |last4=Lee |first4=Chei Sian |date=September 2015 |title=Why Students Share Misinformation on Social Media: Motivation, Gender, and Study-level Differences |journal=The Journal of Academic Librarianship |volume=41 |issue=5 |pages=583–592 |doi=10.1016/j.acalib.2015.07.003 |s2cid=141523357}}</ref>

Agent-based models and other computational models have been used by researchers to explain how false beliefs spread through networks. ] is one example of a computational method for evaluating connections in data shared in a social media network or similar network.<ref>{{cite journal |last1=Shaffer |first1=David Williamson |last2=Collier |first2=Wesley | last3=Ruis | first3=A. R. |title=A tutorial on epistemic network analysis: Analysing the structural connections in cognitive, social and interaction data |journal=Journal of Learning Analytics |date=2016 |volume=3 |issue=3 |pages=9–45 |id={{ERIC|EJ1126800}} |doi=10.18608/jla.2016.33.3 |doi-access=free }}</ref>

Researchers fear that misinformation in social media is "becoming unstoppable".<ref name="Allcott Gentzkow Yu 2019" /> It has also been observed that misinformation and disinformation reappear on social media sites.{{Citation needed|date=October 2023}}

Misinformation spread by ] has been difficult for social media platforms to address.<ref>{{Cite news|last=Milman|first=Oliver|date=2020-02-21|title=Revealed: quarter of all tweets about climate crisis produced by bots|work=The Guardian|url=https://www.theguardian.com/technology/2020/feb/21/climate-tweets-twitter-bots-analysis|url-status=live|access-date=2020-02-23|archive-url=https://web.archive.org/web/20200222193221/https://www.theguardian.com/technology/2020/feb/21/climate-tweets-twitter-bots-analysis|archive-date=2020-02-22 }}</ref> Sites such as Facebook have algorithms that have been proven to further the spread of misinformation in which how content is spread among subgroups.<ref>{{Cite web |title=Facebook's Design Makes It Unable to Control Misinformation {{!}} GW Today {{!}} The George Washington University |url=https://gwtoday.gwu.edu/facebooks-design-makes-it-unable-control-misinformation |access-date=2024-03-04 |website=GW Today }}</ref>

==== Social causes and echo chambers ====
Spontaneous spread of misinformation on social media usually occurs from users sharing posts from friends or mutually-followed pages.<ref name=":9" /> These posts are often shared from someone the sharer believes they can trust.<ref name=":9" /> Misinformation introduced through a social format influences individuals drastically more than misinformation delivered non-socially.<ref>{{cite journal |last1=Gabbert |first1=Fiona |last2=Memon |first2=Amina |last3=Allan |first3=Kevin |last4=Wright |first4=Daniel B. |title=Say it to my face: Examining the effects of socially encountered misinformation |journal=Legal and Criminological Psychology |date=September 2004 |volume=9 |issue=2 |pages=215–227 |doi=10.1348/1355325041719428 |s2cid=144823646 |url=https://rke.abertay.ac.uk/en/publications/c84006a6-18c9-4aec-b44d-da1ec1a8c110 }}</ref>

People are inclined to follow or support like-minded individuals, creating ] and ]s.<ref name=":10" /> Untruths or general agreement within isolated social clusters are difficult to counter.<ref name=":10" /> Some argue this causes an absence of a collective reality.<ref name=":10">{{Cite web|url=https://www.cjr.org/analysis/breitbart-media-trump-harvard-study.php|title=Study: Breitbart-led rightwing media ecosystem altered broader media agenda|last=Benkler|first=Y.|date=2017|access-date=8 June 2018|archive-url=https://web.archive.org/web/20180604140114/https://www.cjr.org/analysis/breitbart-media-trump-harvard-study.php|archive-date=4 June 2018|url-status=live}}</ref> Research has also shown that viral misinformation may spread more widely as a result of echo chambers, as the echo chambers provide an initial seed which can fuel broader viral diffusion.<ref>{{cite journal |last1=Törnberg |first1=P. |year=2018 |title=Echo chambers and viral misinformation: Modeling fake news as complex contagion |journal=PLOS ONE |volume=13 |issue=9 |pages=e0203958 |doi=10.1371/journal.pone.0203958 |doi-access=free |pmid=30235239 |pmc=6147442 |bibcode=2018PLoSO..1303958T }}</ref>

Misinformation might be created and spread with malicious intent for reasons such as causing anxiety or deceiving audiences.<ref name=":9">{{Cite book|last1=Thai|first1=My T.|url=https://books.google.com/books?id=CA4NDgAAQBAJ&q=causes+of+misinformation&pg=PA125|title=Big Data in Complex and Social Networks|last2=Wu|first2=Weili|author2-link=Weili Wu|last3=Xiong|first3=Hui|date=2016-12-01|publisher=CRC Press|isbn=978-1-315-39669-9}}</ref> Rumors created with or without malicious intent may be unknowingly shared by users.{{Citation needed|date=October 2023}} People may know what the scientific community has proved as a fact, and still refuse to accept it as such.<ref>{{cite journal |last1=Scheufele |first1=Dietram A. |last2=Krause |first2=Nicole M. |title=Science audiences, misinformation, and fake news |journal=Proceedings of the National Academy of Sciences |date=16 April 2019 |volume=116 |issue=16 |pages=7662–7669 |doi=10.1073/pnas.1805871115 |pmid=30642953 |pmc=6475373 |bibcode=2019PNAS..116.7662S |doi-access=free }}</ref>

==== Lack of regulation ====
Misinformation on social media spreads quickly in comparison to ] because of the lack of regulation and examination required before posting.<ref name="TJAL" /><ref>{{cite book |last1=Caramancion |first1=Kevin Matthe |title=Computational Data and Social Networks |year=2021 |isbn=978-3-030-91433-2 |series=Lecture Notes in Computer Science |volume=13116 |pages=319–329 |chapter=The Role of Information Organization and Knowledge Structuring in Combatting Misinformation: A Literary Analysis |doi=10.1007/978-3-030-91434-9_28 |s2cid=244890285}}</ref>

Social media sites provide users with the capability to spread information quickly to other users without requiring the permission of a gatekeeper such as an editor, who might otherwise require confirmation of the truth before allowing publication.<ref name=":7">{{cite book |last1=Starbird |first1=Kate |title=Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems |last2=Dailey |first2=Dharma |last3=Mohamed |first3=Owla |last4=Lee |first4=Gina |last5=Spiro |first5=Emma S. |date=19 April 2018 |isbn=978-1-4503-5620-6 |pages=1–12 |chapter=Engage Early, Correct More: How Journalists Participate in False Rumors Online during Crisis Events |doi=10.1145/3173574.3173679 |s2cid=5046314}}</ref><ref name=":8">{{cite book |last1=Arif |first1=Ahmer |title=Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing |last2=Robinson |first2=John J. |last3=Stanek |first3=Stephanie A. |last4=Fichet |first4=Elodie S. |last5=Townsend |first5=Paul |last6=Worku |first6=Zena |last7=Starbird |first7=Kate |year=2017 |isbn=978-1-4503-4335-0 |pages=155–168 |chapter=A Closer Look at the Self-Correcting Crowd |doi=10.1145/2998181.2998294 |s2cid=15167363}}</ref>

The problem of misinformation in social media is getting worse as younger generations prefer social media over journalistic for their source of information.<ref>{{Cite journal |last1=Spradling |first1=Matthew |last2=Straub |first2=Jeremy |last3=Strong |first3=Jay |date=June 2021 |title=Protection from 'Fake News': The Need for Descriptive Factual Labeling for Online Content |journal=Future Internet |volume=13 |issue=6 |pages=142 |doi=10.3390/fi13060142 |issn=1999-5903 |doi-access=free}}</ref>

==== Countermeasures ====
Combating the spread of misinformation on social medias is difficult for reasons such as :

* the profusion of misinformation sources makes the reader's task of weighing the reliability of information more challenging<ref>{{cite web |last1=Messerole |first1=Chris |date=2018-05-09 |title=How misinformation spreads on social media&nbsp;– And what to do about it |url=https://www.brookings.edu/blog/order-from-chaos/2018/05/09/how-misinformation-spreads-on-social-media-and-what-to-do-about-it/ |url-status=live |archive-url=https://web.archive.org/web/20190225044815/https://www.brookings.edu/blog/order-from-chaos/2018/05/09/how-misinformation-spreads-on-social-media-and-what-to-do-about-it/ |archive-date=25 February 2019 |access-date=24 February 2019 |website=Brookings Institution}}</ref>
* social media's propensity for ]s embeds misinformation with identity-based conflict<ref name=":4">{{Cite journal |last1=Diaz Ruiz |first1=Carlos |last2=Nilsson |first2=Tomas |date=2022-08-08 |title=Disinformation and Echo Chambers: How Disinformation Circulates on Social Media Through Identity-Driven Controversies |journal=Journal of Public Policy & Marketing |volume=42 |pages=18–35 |doi=10.1177/07439156221103852 |issn=0743-9156|doi-access=free }}</ref>
* the proliferation of ] form an epistemic environment in which participants encounter beliefs and opinions that coincide with their own,<ref name=":5">{{Cite journal |last=Nguyen |first=C. Thi |title=Echo Chambers and Epistemic Bubbles |date=2020 |url=https://www.cambridge.org/core/journals/episteme/article/abs/echo-chambers-and-epistemic-bubbles/5D4AC3A808C538E17C50A7C09EC706F0 |journal=Episteme |volume=17 |issue=2 |pages=141–161 |doi=10.1017/epi.2018.32 |s2cid=171520109 |issn=1742-3600 |access-date=2022-11-25 |archive-date=2022-02-09 |archive-url=https://web.archive.org/web/20220209135355/https://www.cambridge.org/core/journals/episteme/article/abs/echo-chambers-and-epistemic-bubbles/5D4AC3A808C538E17C50A7C09EC706F0 |url-status=live }}</ref> moving the entire group toward more extreme positions.<ref name=":5" /><ref name=":4" />

With the large audiences that can be reached and the experts on various subjects on social media, some believe social media could also be the key to correcting misinformation.<ref name="HealthComm">{{cite journal |last1=Bode |first1=Leticia |last2=Vraga |first2=Emily K. |title=See Something, Say Something: Correction of Global Health Misinformation on Social Media |journal=Health Communication |date=2 September 2018 |volume=33 |issue=9 |pages=1131–1140 |doi=10.1080/10410236.2017.1331312 |pmid=28622038 |s2cid=205698884 }}</ref>

Journalists today are criticized for helping to spread false information on these social platforms, but research shows they also play a role in curbing it through debunking and denying false rumors.<ref name=":7" /><ref name=":8" />

==== COVID-19 misinformation ====
{{Main|COVID-19 misinformation}}
During the ], social media was used as one of the main propagators for spreading ] about symptoms, treatments, and long-term health-related problems.<ref name=":2" /> This problem has initialized a significant effort in developing automated detection methods for misinformation on social media platforms.<ref name=":3" />

The creator of the ] made money posting anti-vax false news on social media. He posted more than 150 posts aimed towards women, garnering a total of 1.6&nbsp;million views and earning money for every click and share.{{sfn|Aral|2020|p=43}}

==== Misinformation on TikTok ====
A research report by ] found there is a very high level (~20% in their probes of videos about relevant topics) of online misinformation delivered – to a mainly young user base – with ], whose (essentially unregulated) usage is increasing as of 2022.<ref>{{cite news |last1=Tucker |first1=Emma |date=18 September 2022 |title=TikTok's search engine repeatedly delivers misinformation to its majority-young user base, report says {{!}} CNN Business |work=CNN |url=https://edition.cnn.com/2022/09/18/business/tiktok-search-engine-misinformation/index.html |url-status=live |access-date=19 October 2022 |archive-url=https://web.archive.org/web/20221019151032/https://edition.cnn.com/2022/09/18/business/tiktok-search-engine-misinformation/index.html |archive-date=19 October 2022}}</ref><ref>{{cite web |title=Misinformation Monitor: September 2022 |url=https://www.newsguardtech.com/misinformation-monitor/september-2022/ |url-status=live |archive-url=https://web.archive.org/web/20221018190939/https://www.newsguardtech.com/misinformation-monitor/september-2022/ |archive-date=18 October 2022 |access-date=19 October 2022 |website=NewsGuard}}</ref>

==== Misinformation on Facebook ====
{{Main|Misinformation on Facebook}}
A research study of Facebook found that misinformation was more likely to be clicked on than factual information.<ref>{{Cite news |last=Dwoskin |first=Elizabeth |title=Misinformation on Facebook got six times more clicks than factual news during the 2020 election, study says |newspaper=The Washington Post |url=https://www.washingtonpost.com/technology/2021/09/03/facebook-misinformation-nyu-study/ |url-status=live |access-date=2021-10-22 |archive-url=https://web.archive.org/web/20211021031124/https://www.washingtonpost.com/technology/2021/09/03/facebook-misinformation-nyu-study/ |archive-date=2021-10-21}}</ref> The most common reasons that Facebook users were sharing misinformation for socially-motivated reasons, rather than taking the information seriously.<ref name="ACM/IEEE-CE">{{cite book |last1=Chen |first1=Xinran |title=Proceedings of the 15th ACM/IEEE-CS Joint Conference on Digital Libraries |last2=Sin |first2=Sei-Ching Joanna |last3=Theng |first3=Yin-Leng |last4=Lee |first4=Chei Sian |year=2015 |isbn=978-1-4503-3594-2 |pages=111–114 |chapter=Why do Social Media Users Share Misinformation? |doi=10.1145/2756406.2756941 |s2cid=15983217}}</ref>

]'s coverage of misinformation has become a ], as some reports indicated Facebook recommended pages containing health misinformation.<ref name="Griffith20210721">{{cite news |last1=Griffith |first1=Chris |date=21 July 2021 |title=Facebook exposed over its handling of COVID misinformation |newspaper=The Australian |location=Canberra |id={{ProQuest|2553642687}}}}</ref> For example, this can be seen when a user likes an ] Facebook page. Automatically, more and more anti-vax pages are recommended to the user.<ref name="Griffith20210721" /> Additionally, some reference Facebook's inconsistent censorship of misinformation leading to deaths from COVID-19.<ref name="Griffith20210721" />

Facebook estimated the existence of up to 60&nbsp;million ] ] actively spreading misinformation on their platform,<ref>{{cite journal |last1=Iyengar |first1=Shanto |last2=Massey |first2=Douglas S. |date=16 April 2019 |title=Scientific communication in a post-truth society |journal=Proceedings of the National Academy of Sciences |volume=116 |issue=16 |pages=7656–7661 |bibcode=2019PNAS..116.7656I |doi=10.1073/pnas.1805868115 |pmc=6475392 |pmid=30478050 |doi-access=free}}</ref> and has taken measures to stop the spread of misinformation, resulting in a decrease, though misinformation continues to exist on the platform.<ref name="Allcott Gentzkow Yu 2019" /> On Facebook, adults older than 65 were seven times more likely to share fake news than adults ages 18–29.<ref name="Swire">{{cite journal |last1=Swire-Thompson |first1=Briony |last2=Lazer |first2=David |date=2 April 2020 |title=Public Health and Online Misinformation: Challenges and Recommendations |journal=Annual Review of Public Health |volume=41 |issue=1 |pages=433–451 |doi=10.1146/annurev-publhealth-040119-094127 |pmid=31874069 |doi-access=free}}</ref>

==== Misinformation on Twitter ====
{{Main|Twitter under Elon Musk#Misinformation and disinformation|Community Notes}}

] is one of the most concentrated platforms for engagement with political ]. 80% of fake news sources are shared by 0.1% of users, who are "super-sharers". Older, more conservative social users are also more likely to interact with fake news.<ref name="ACM/IEEE-CE" /> Another source of misinformation on Twitter are ], especially surrounding ].<ref>{{Cite web|date=2020-02-21|title=Revealed: a quarter of all tweets about climate crisis produced by bots|url=http://www.theguardian.com/technology/2020/feb/21/climate-tweets-twitter-bots-analysis|access-date=2021-04-20|website=The Guardian|archive-date=2021-04-29|archive-url=https://web.archive.org/web/20210429020205/https://www.theguardian.com/technology/2020/feb/21/climate-tweets-twitter-bots-analysis|url-status=live}}</ref> Bot accounts on Twitter accelerate true and fake news at the same rate.<ref>{{Cite journal |last1=Vosoughi |first1=Soroush |last2=Roy |first2=Deb |last3=Aral |first3=Sinan |date=2018-03-09 |title=The spread of true and false news online |journal=Science |volume=359 |issue=6380 |pages=1146–1151 |doi=10.1126/science.aap9559 |pmid=29590045 |bibcode=2018Sci...359.1146V |s2cid=4549072 |issn=0036-8075 |doi-access=free }}</ref> A 2018 study of Twitter determined that, compared to accurate information, false information spread significantly faster, further, deeper, and more broadly.<ref name="Swire" /> A research study watched the process of thirteen rumors appearing on Twitter and noticed that eleven of those same stories resurfaced multiple times, after time had passed.<ref>{{cite journal |last1=Shin |first1=Jieun |last2=Jian |first2=Lian |last3=Driscoll |first3=Kevin |last4=Bar |first4=François |date=June 2018 |title=The diffusion of misinformation on social media: Temporal pattern, message, and source |journal=Computers in Human Behavior |volume=83 |pages=278–287 |doi=10.1016/j.chb.2018.02.008 |s2cid=41956979}}</ref>

A social media app called ] has caused much chaos as well. Right winged Twitter users who were banned on the app moved to Parler after the ], and the app was being used to plan and facilitate more illegal and dangerous activities. Google and Apple later pulled the app off their respective app stores. This app has been able to cause a lot of misinformation and bias in the media, allowing for more political mishaps.<ref>{{cite news |date=10 January 2021 |title=Amazon to suspend Parler after deadly Capitol Hill riot |work=Al Jazeera |url=https://www.aljazeera.com/news/2021/1/10/amazon-to-suspend-parler-after-deadly-capitol-hill-riot |url-status=live |access-date=7 March 2022 |archive-url=https://web.archive.org/web/20220307030229/https://www.aljazeera.com/news/2021/1/10/amazon-to-suspend-parler-after-deadly-capitol-hill-riot |archive-date=7 March 2022}}</ref>

==== Misinformation on YouTube ====
{{excerpt|Social impact of YouTube|False scientific content}}

=== Lack of peer review ===
]
Due to the decentralized nature and structure of the Internet, content creators can easily publish content without being required to undergo ], prove their qualifications, or provide backup documentation. While library books have generally been reviewed and edited by an editor, publishing company, etc., Internet sources cannot be assumed to be vetted by anyone other than their authors. Misinformation may be produced, reproduced, and posted immediately on most online platforms.<ref>{{cite journal | last=Stapleton | first=Paul | title=Assessing the quality and bias of web-based sources: implications for academic writing | journal=Journal of English for Academic Purposes | year=2003 | volume=2 | issue=3 | pages=229–245 | doi=10.1016/S1475-1585(03)00026-2}}</ref><ref>{{cite journal |last1=West |first1=Jevin D. |last2=Bergstrom |first2=Carl T. |title=Misinformation in and about science |journal=] |date=2021 |volume=118 |issue=15 |doi=10.1073/pnas.1912444117 |pmid=33837146 |pmc=8054004 |bibcode=2021PNAS..11812444W |doi-access=free}}</ref>

=== Censorship accusations ===
Social media sites such as Facebook and Twitter have found themselves defending accusations of ] for removing posts they have deemed to be misinformation. Social media censorship policies relying on government agency-issued guidance to determine information validity have garnered criticism that such policies have the unintended effect of stifling dissent and criticism of government positions and policies.<ref name=":0">{{Cite news|title=Facebook's Lab-Leak About-Face|newspaper=The Wall Street Journal|url=https://www.wsj.com/articles/facebooks-lab-leak-about-face-11622154198|url-status=live|access-date=2021-07-14|archive-date=2021-07-14|archive-url=https://web.archive.org/web/20210714010553/https://www.wsj.com/articles/facebooks-lab-leak-about-face-11622154198}}</ref> Most recently, social media companies have faced criticism over allegedly prematurely censoring the discussion of the ].<ref name=":0" /><ref>{{Cite news|title=Covid origin: Why the Wuhan lab-leak theory is being taken seriously|work=BBC News|date=27 May 2021|url=https://www.bbc.com/news/world-asia-china-57268111|url-status=live|access-date=14 July 2021|archive-date=30 June 2021|archive-url=https://web.archive.org/web/20210630155548/https://www.bbc.com/news/world-asia-china-57268111}}</ref>

Other accusations of censorship appear to stem from attempts to prevent social media consumers from self-harm through the use of unproven COVID-19 treatments. For example, in July 2020, a video went viral showing Dr. Stella Immanuel claiming ] was an effective cure for COVID-19. In the video, Immanuel suggested that there was no need for masks, school closures, or any kind of economic shut down; attesting that her alleged cure was highly effective in treating those infected with the virus. The video was shared 600,000 times and received nearly 20 million views on Facebook before it was taken down for violating community guidelines on spreading misinformation.<ref name=":1">{{Cite news|date=2020-07-28|title=Hydroxychloroquine: Why a video promoted by Trump was pulled on social media|work=BBC News|url=https://www.bbc.com/news/53559938|access-date=2021-11-24|archive-date=2020-11-02|archive-url=https://web.archive.org/web/20201102215305/https://www.bbc.com/news/53559938|url-status=live}}</ref> The video was also taken down on Twitter overnight, but not before former president ] shared it to his page, which was followed by over 85 million Twitter users. ] director Dr. Anthony Fauci and members of the ] (WHO) quickly discredited the video, citing larger-scale studies of hydroxychloroquine showing it is not an effective treatment of COVID-19, and the ] cautioned against using it to treat COVID-19 patients following evidence of serious heart problems arising in patients who have taken the drug.<ref name=BBCNews>{{Cite news|date=2020-07-29|title=Stella Immanuel – the doctor behind unproven coronavirus cure claim|work=BBC News|url=https://www.bbc.com/news/world-africa-53579773|access-date=2020-11-23|archive-date=2021-10-11|archive-url=https://web.archive.org/web/20211011045822/https://www.bbc.com/news/world-africa-53579773|url-status=live}}</ref>

Another prominent example of misinformation removal criticized by some as an example of censorship was the '']''{{'s}} report on the ] laptops approximately two weeks before the 2020 presidential election, which was used to promote the ]. Social media companies quickly removed this report, and the ''Post'''s Twitter account was temporarily suspended. Over 50 intelligence officials found the disclosure of emails allegedly belonging to ]'s son had all the "classic earmarks of a Russian information operation".<ref name="officials">{{Cite web|last=Bertrand|first=Natasha|author-link=Natasha Bertrand|date=October 19, 2020|title=Hunter Biden story is Russian disinfo, dozens of former intel officials say|url=https://www.politico.com/news/2020/10/19/hunter-biden-story-russian-disinfo-430276|url-status=live|archive-url=https://web.archive.org/web/20201020034222/https://www.politico.com/news/2020/10/19/hunter-biden-story-russian-disinfo-430276|archive-date=October 20, 2020|access-date=October 20, 2020|website=]}}</ref> Later evidence emerged that at least some of the laptop's contents were authentic.<ref name="laptop">{{cite web|last=Lizza|first=Ryan|author-link=Ryan Lizza|date=September 21, 2021|title=Politico Playbook: Double Trouble for Biden|url=https://www.politico.com/newsletters/playbook/2021/09/21/double-trouble-for-biden-494411|website=]|access-date=September 23, 2021|archive-date=November 23, 2021|archive-url=https://web.archive.org/web/20211123034707/https://www.politico.com/newsletters/playbook/2021/09/21/double-trouble-for-biden-494411|url-status=live}}</ref>

== Noteworthy examples ==
An example of bad information from media sources that led to the spread of misinformation occurred in November 2005, when ] on '']'' claimed that law enforcement officials estimate 50,000 predators are online at any moment. Afterward, the U.S. attorney general at the time, ], repeated the claim. However, the number that Hansen used in his reporting had no backing. Hansen said he received the information from ''Dateline'' expert Ken Lanning, but Lanning admitted that he made up the number 50,000 because there was no solid data on the number. According to Lanning, he used 50,000 because it sounds like a real number, not too big and not too small, and referred to it as a "] number". Reporter Carl Bialik says that the number 50,000 is used often in the media to estimate numbers when reporters are unsure of the exact data.<ref>{{Cite book|title=The Influencing Machine|last=Gladstone|first=Brooke|publisher=W. W. Norton & Company|year=2012|isbn=978-0-393-34246-8|location=New York|pages=49–51}}</ref>

During the ], a ] that ] gained significant traction worldwide after emerging on social media.<ref name="Stewart 2021 106–110">{{cite journal |last1=Stewart |first1=Mallory |date=2021 |title=Defending Weapons Inspections from the Effects of Disinformation |journal=AJIL Unbound |volume=115 |pages=106–110 |doi=10.1017/aju.2021.4 |s2cid=232070073 |doi-access=free}}</ref>

Misinformation was a major talking point during the ] with claims of social media sites allowing "]" to be spread.<ref name="Brosnan20210113">{{cite news |id={{ProQuest|2477885938}} |last1=Brosnan |first1=Deanne |date=13 January 2021 |title=When Misinformation is Misinformation |newspaper=CE Think Tank Newswire |location=Miami }}</ref>

== Impact ==
The ] describes a situation in which individuals are so concerned about realistic misinformation (in particular, ]s) that they begin to mistrust real content, particularly if someone claims that it is false.<ref name=":15">{{Cite web |date=2024-02-08 |title=Deepfakes, Elections, and Shrinking the Liar's Dividend {{!}} Brennan Center for Justice |url=https://www.brennancenter.org/our-work/research-reports/deepfakes-elections-and-shrinking-liars-dividend |access-date=2024-04-12 |website=www.brennancenter.org}}</ref> For instance, a politician could benefit from claiming that a real video of them doing something embarrassing was actually AI-generated or altered, leading followers to mistrust something that was actually real. On a larger scale this problem can lead to erosion in the public's trust of generally reliable information sources.<ref name=":15" />

Misinformation can affect all aspects of life. Allcott, Gentzkow, and Yu concur that the diffusion of misinformation through social media is a potential threat to democracy and broader society. The effects of misinformation can lead to decline of accuracy of information as well as event details.<ref>{{cite journal |last1=Bodner |first1=Glen E. |last2=Musch |first2=Elisabeth |last3=Azad |first3=Tanjeem |title=Reevaluating the potency of the memory conformity effect |journal=Memory & Cognition |date=December 2009 |volume=37 |issue=8 |pages=1069–1076 |doi=10.3758/MC.37.8.1069 |pmid=19933452 |doi-access=free }}</ref> When eavesdropping on conversations, one can gather facts that may not always be true, or the receiver may hear the message incorrectly and spread the information to others. On the Internet, one can read content that is stated to be factual but that may not have been checked or may be erroneous. In the news, companies may emphasize the speed at which they receive and send information but may not always be correct in the facts. These developments contribute to the way misinformation may continue to complicate the public's understanding of issues and to serve as a source for belief and attitude formation.<ref>{{Cite book|url=https://books.google.com/books?id=1Jo7DwAAQBAJ&q=misinformation+causes&pg=PT246|title=Misinformation and Mass Audiences|last1=Southwell|first1=Brian G.|last2=Thorson|first2=Emily A.|last3=Sheble|first3=Laura|date=2018|publisher=University of Texas Press|isbn=978-1-4773-1458-6}}</ref>

In regards to politics, some view being a misinformed citizen as worse than being an uninformed citizen. Misinformed citizens can state their beliefs and opinions with confidence and thus affect elections and policies. This type of misinformation occurs when a speaker appears "authoritative and legitimate", while also spreading misinformation.<ref name="The Growing Role of Social Media in" /> When information is presented as vague, ambiguous, sarcastic, or partial, receivers are forced to piece the information together and make assumptions about what is correct.<ref>{{cite book | last=Barker | first=David | title=Rushed to Judgement: Talk Radio, Persuasion, and American Political Behavior | year=2002 | publisher=Columbia University Press | location=New York | pages=106–109}}</ref> Misinformation has the power to sway public elections and referendums if it gains enough momentum. Leading up to the ], for example, a figure used prominently by the ] campaign claimed that by leaving the EU the UK would save £350 million a week, 'for the NHS'. Claims then circulated widely in the campaign that this amount ''would'' (rather than ''could'' theoretically) be redistributed to the British ] after Brexit. This was later deemed a "clear misuse of official statistics" by the UK statistics authority.

Moreover, the advert infamously shown on the side of London's double-decker busses did not take into account the UK's budget rebate, and the idea that 100% of the money saved would go to the NHS was unrealistic. A poll published in 2016 by ] found that nearly half of the British public believed this misinformation to be true.<ref>{{cite news |title=The misinformation that was told about Brexit during and after the referendum |url=https://www.independent.co.uk/news/uk/politics/final-say-brexit-referendum-lies-boris-johnson-leave-campaign-remain-a8466751.html |work=The Independent |date=2 August 2018 |archive-url=https://ghostarchive.org/archive/20220515/https://www.independent.co.uk/news/uk/politics/final-say-brexit-referendum-lies-boris-johnson-leave-campaign-remain-a8466751.html |archive-date=15 May 2022 }}</ref> Even when information is proven to be misinformation, it may continue to shape attitudes towards a given topic,<ref name="MediaMisinfo">{{Cite book |last1=Marwick |first1=Alice |title=Media Manipulation and Disinformation Online |last2=Lewis |first2=Rebecca |publisher=Data & Society Research Institute |year=2017 |location=New York |pages=40–45}}</ref> meaning it has the power to swing political decisions if it gains enough traction. A study conducted by Soroush Vosoughi, Deb Roy and Sinan Aral looked at Twitter data including 126,000 posts spread by 3 million people over 4.5 million times. They found that political news traveled faster than any other type of information. They found that false news about politics reached more than 20,000 people three times faster than all other types of false news.{{sfn|Aral|2020|pp=45-47}}

Aside from ], misinformation can also be employed in industrial propaganda. Using tools such as advertising, a company can undermine reliable evidence or influence belief through a concerted misinformation campaign. For instance, tobacco companies employed misinformation in the second half of the twentieth century to diminish the reliability of studies that demonstrated the link between ] and ].<ref>{{Cite book|last1=O'Connor|first1=Cailin|url=https://archive.org/details/misinformationag0000ocon|title=The Misinformation Age: How False Beliefs Spread|last2=Weatherall|first2=James Owen|publisher=Yale University Press|year=2019|isbn=978-0-300-23401-5|location=New Haven|pages=|url-access=limited}}</ref>

In the medical field, misinformation can immediately lead to life endangerment as seen in the case of the public's negative perception towards vaccines or the use of herbs instead of medicines to treat diseases.<ref name="The Growing Role of Social Media in" /><ref>{{Cite book|last1=Sinha|first1=P.|url=https://books.google.com/books?id=9ZuMDwAAQBAJ&q=misinformation+causes&pg=PT22|title=India Misinformed: The True Story|last2=Shaikh|first2=S.|last3=Sidharth|first3=A.|publisher=Harper Collins|year=2019|isbn=978-93-5302-838-1}}</ref> In regards to the ], the spread of misinformation has proven to cause confusion as well as negative emotions such as anxiety and fear.<ref name="Bratu 2020">{{cite journal|last1=Bratu|first1=Sofia|date=May 24, 2020|title=The Fake News Sociology of COVID-19 Pandemic Fear: Dangerously Inaccurate Beliefs, Emotional Contagion, and Conspiracy Ideation|journal=Linguistic and Philosophical Investigations|volume=19|pages=128–134|doi=10.22381/LPI19202010|doi-access=free}}</ref><ref>{{Cite Q|Q97652640}}</ref> Misinformation regarding proper safety measures for the prevention of the virus that go against information from legitimate institutions like the ] can also lead to inadequate protection and possibly place individuals at risk for exposure.<ref name="Bratu 2020" /><ref>{{Cite web|date=2020-07-29|title=Misinformation on coronavirus is proving highly contagious|url=https://apnews.com/article/ap-top-news-understanding-the-outbreak-health-media-social-media-86f61f3ffb6173c29bc7db201c10f141|access-date=2020-11-23|website=AP NEWS|archive-date=2020-11-20|archive-url=https://web.archive.org/web/20201120041927/https://apnews.com/article/ap-top-news-understanding-the-outbreak-health-media-social-media-86f61f3ffb6173c29bc7db201c10f141|url-status=live}}</ref>

Some scholars and activists are heading movements to eliminate the mis/disinformation and information pollution in the digital world. One theory, "information environmentalism," has become a curriculum in some universities and colleges.<ref>{{Cite news|url=https://er.educause.edu/articles/2017/10/info-environmentalism-an-introduction|title=Info-Environmentalism: An Introduction|access-date=2018-09-28|archive-url=https://web.archive.org/web/20180703130624/https://er.educause.edu/articles/2017/10/info-environmentalism-an-introduction|archive-date=2018-07-03|url-status=live}}</ref><ref>{{Cite news|url=https://dlinq.middcreate.net/informationenvironmentalism/|title=Information Environmentalism|date=2017-12-21|work=Digital Learning and Inquiry (DLINQ)|access-date=2018-09-28|archive-url=https://web.archive.org/web/20180928044347/https://dlinq.middcreate.net/informationenvironmentalism/|archive-date=2018-09-28|url-status=live}}</ref> The general study of misinformation and disinformation is by now also common across various academic disciplines, including sociology, communication, computer science, and political science, leading to the emerging field being described loosely as "]".<ref>{{cite journal |last1=Righetti |first1=Nicola |last2=Rossi |first2=Luca |last3=Marino |first3=Giada |title=At the onset of an infodemic: Geographic and disciplinary boundaries in researching problematic COVID-19 information |journal=First Monday |date=4 July 2022 |doi=10.5210/fm.v27i7.12557 |s2cid=250289817 |doi-access=free }}</ref> However, various scholars and journalists have criticised this development, pointing to problematic normative assumptions, a varying quality of output and lack of methodological rigor, as well as a too strong impact of mis- and disinformation research in shaping public opinion and policymaking.<ref>{{cite magazine |last1=Bernstein |first1=Joseph |title=Bad News: Selling the story of disinformation |url=https://harpers.org/archive/2021/09/bad-news-selling-the-story-of-disinformation/ |magazine=Harper's Magazine |date=9 August 2021 |access-date=30 September 2022 |archive-date=26 September 2022 |archive-url=https://web.archive.org/web/20220926211045/https://harpers.org/archive/2021/09/bad-news-selling-the-story-of-disinformation/ |url-status=live }}</ref><ref>{{Cite web |last=Adler-Bell |first=Sam |date=2022-05-20 |title=The Liberal Obsession With 'Disinformation' Is Not Helping |url=https://nymag.com/intelligencer/2022/05/the-liberal-obsession-with-disinformation-is-not-helping.html |access-date=2022-09-30 |website=Intelligencer |archive-date=2022-09-30 |archive-url=https://web.archive.org/web/20220930145425/https://nymag.com/intelligencer/2022/05/the-liberal-obsession-with-disinformation-is-not-helping.html |url-status=live }}</ref> Summarising the most frequent points of critique, communication scholars Chico Camargo and Felix Simon wrote in an article for the Harvard Kennedy School Misinformation Review that "mis-/disinformation studies has been accused of lacking clear definitions, having a simplified understanding of what it studies, a too great emphasis on media effects, a neglect of intersectional factors, an outsized influence of funding bodies and policymakers on the research agenda of the field, and an outsized impact of the field on policy and policymaking."<ref>{{cite journal |last1=Camargo |first1=Chico Q. |last2=Simon |first2=Felix M. |title=Mis- and disinformation studies are too big to fail: Six suggestions for the field's future |journal=Harvard Kennedy School Misinformation Review |date=20 September 2022 |doi=10.37016/mr-2020-106 |s2cid=252423678 |doi-access=free }}</ref>

== AI contribution to the problem and aid in combatting ==
] exacerbates the problem of misinformation but also contributes to the fight against misinformation.

=== AI's contribution to the problem ===
* ]s and ] create very convincing visual, audio, and textual evidence that is difficult to distinguish from legitimate authoritative evidence.<ref>{{Cite journal |last=Westerlund |first=Mika |date=2019-01-01 |title=The Emergence of Deepfake Technology: A Review |journal=Technology Innovation Management Review |volume=9 |issue=11 |pages=39–52 |doi=10.22215/timreview/1282 |issn=1927-0321|doi-access=free }}</ref>
* ]s and automated ]s can rapidly sow disinformation.<ref>{{Cite journal |last=Ferrara |first=Emilio |date=2017-07-31 |title=Disinformation and Social Bot Operations in the Run Up to the 2017 French Presidential Election |journal=First Monday |doi=10.5210/fm.v22i8.8005 |doi-access=free |arxiv=1707.00086 |issn=1396-0466}}</ref>
* ] plays a role in amplification of sensational and controversial material regardless of truth.<ref>{{Cite journal |last1=Bakshy |first1=Eytan |last2=Messing |first2=Solomon |last3=Adamic |first3=Lada A. |date=2015-06-05 |title=Exposure to ideologically diverse news and opinion on Facebook |url=http://dx.doi.org/10.1126/science.aaa1160 |journal=Science |volume=348 |issue=6239 |pages=1130–1132 |doi=10.1126/science.aaa1160 |pmid=25953820 |bibcode=2015Sci...348.1130B |issn=0036-8075}}</ref>

=== AI as a tool to combat misinformation ===
* ] algorithms are employed to fact-check truth claims in real-time.<ref>{{Cite web |title=Document-level Claim Extraction and Decontextualisation for Fact-Checking |url=https://arxiv.org/html/2406.03239v2 |access-date=2024-10-10 |website=arxiv.org}}</ref>
* Researchers are developing AI tools for detecting fabricated audio and video.<ref>{{Cite book |last1=Afchar |first1=Darius |last2=Nozick |first2=Vincent |last3=Yamagishi |first3=Junichi |last4=Echizen |first4=Isao |chapter=MesoNet: A Compact Facial Video Forgery Detection Network |date=December 2018 |title=2018 IEEE International Workshop on Information Forensics and Security (WIFS) |chapter-url=http://dx.doi.org/10.1109/wifs.2018.8630761 |pages=1–7 |publisher=IEEE |doi=10.1109/wifs.2018.8630761|arxiv=1809.00888 |isbn=978-1-5386-6536-7 }}</ref>
* AI can be used for ] and ] education.<ref>{{Cite journal |last1=Carroll |first1=Alexander J. |last2=Borycz |first2=Joshua |date=July 2024 |title=Integrating large language models and generative artificial intelligence tools into information literacy instruction |journal=The Journal of Academic Librarianship |language=en |volume=50 |issue=4 |pages=102899 |doi=10.1016/j.acalib.2024.102899|doi-access=free }}</ref>

== See also ==
{{columns-list|colwidth=20em|
* ] * ]
* ]
* ] * ]
* ] * ]
* ]
* {{Annotated link|Artificial intelligence and elections}}
* ]
* ] * ]
** ] (also known as "slander") ** ] (also known as "slander")
* ] * ]
* ] * ]
* ]
* ] * ]
* ] * ]
** ] ** ]
* ]
* ] * ]
* ] * ]
* ] * ]
* ] * ]
* ]
* ]
* ]
* ]
* ]
* ] * ]
* ]
* ] * ]
* ]
* Social engineering (in ] and ]) * Social engineering (in ] and ])
* ]
* ]
* ]
}} }}


==References== == References ==
{{reflist}}
<references />


==Further reading== == Further reading ==
{{Library resources box}} {{Library resources box}}
* {{cite book |doi=10.1145/3308560.3316738 |chapter=A Study of Misinformation in WhatsApp groups with a focus on the Brazilian Presidential Elections |title=Companion Proceedings of the 2019 World Wide Web Conference |year=2019 |last1=Machado |first1=Caio |last2=Kira |first2=Beatriz |last3=Narayanan |first3=Vidya |last4=Kollanyi |first4=Bence |last5=Howard |first5=Philip |pages=1013–1019 |isbn=978-1-4503-6675-5 |s2cid=153314118 }}

* {{Cite journal|last1=Machado|first1=Caio|last2=Kira|first2=Beatriz|last3=Narayanan|first3=Vidya|last4=Kollanyi|first4=Bence|last5=Howard|first5=Philip|s2cid=153314118|date=2019|title=A Study of Misinformation in WhatsApp groups with a focus on the Brazilian Presidential Elections.|journal=Companion Proceedings of the 2019 World Wide Web Conference on WWW '19|pages=1013–1019|location=New York|publisher=ACM Press|doi=10.1145/3308560.3316738|isbn=978-1450366755}} * {{cite journal | last1 = Allcott | first1 = H. | last2 = Gentzkow | first2 = M. | s2cid = 32730475 | year = 2017 | title = Social Media and Fake News in the 2016 Election | journal = Journal of Economic Perspectives | volume = 31 | issue = 2| pages = 211–236 | doi = 10.1257/jep.31.2.211 | doi-access = free }}
* {{Cite book |last=Aral |first=Sinan |title=The Hype Machine: How Social Media Disrupts Our Elections, Our Economy, and Our Health--and How We Must Adapt |publisher=Currency |date=2020 |isbn=978-0-525-57451-4 |edition=First |location=New York |oclc=1155486056}}
* {{cite journal | last1 = Allcott | first1 = H. | last2 = Gentzkow | first2 = M. | s2cid = 32730475 | year = 2017 | title = Social Media and Fake News in the 2016 Election | url = | journal = Journal of Economic Perspectives | volume = 31 | issue = 2| pages = 211–236 | doi = 10.1257/jep.31.2.211 | doi-access = free }}
* Baillargeon, Normand (4 January 2008). A short course in intellectual self-defense. Seven Stories Press. {{ISBN|978-1-58322-765-7}}. Retrieved 22 June 2011. * Baillargeon, Normand (4 January 2008). A short course in intellectual self-defense. Seven Stories Press. {{ISBN|978-1-58322-765-7}}. Retrieved 22 June 2011.
* {{cite journal | last1 = Bakir | first1 = V. | last2 = McStay | first2 = A. | s2cid = 157153522 | year = 2017 | title = Fake News and The Economy of Emotions: Problems, causes, solutions | url = https://research.bangor.ac.uk/portal/en/researchoutputs/fake-news-and-the-economy-of-emotions(6f96b5ed-884a-43c1-921f-74ed6f1384f8).html| journal = Digital Journalism | volume = 6| issue = | pages = 154–175| doi = 10.1080/21670811.2017.1345645 }} * {{cite journal |last1=Bakir |first1=Vian |last2=McStay |first2=Andrew |title=Fake News and The Economy of Emotions: Problems, causes, solutions |journal=Digital Journalism |date=7 February 2018 |volume=6 |issue=2 |pages=154–175 |doi=10.1080/21670811.2017.1345645 |s2cid=157153522 |url=https://research.bangor.ac.uk/portal/en/researchoutputs/fake-news-and-the-economy-of-emotions(6f96b5ed-884a-43c1-921f-74ed6f1384f8).html }}
* ], and ], ''The Experts Speak: The Definitive Compendium of Authoritative Misinformation'', Pantheon Books, 1984. * ], and ], ''The Experts Speak: The Definitive Compendium of Authoritative Misinformation'', Pantheon Books, 1984.
* {{cite journal|author=Cook, John|author2= Stephan Lewandowsky|author3=Ullrich K. H. Ecker|title=Neutralizing misinformation through inoculation: Exposing misleading argumentation techniques reduces their influence|journal=]|date=2017-05-05|volume= 12|issue= 5|pages= e0175799|doi=10.1371/journal.pone.0175799|pmid= 28475576|pmc= 5419564|bibcode= 2017PLoSO..1275799C|doi-access=free}} * {{cite journal|author=Cook, John|author2= Stephan Lewandowsky|author3=Ullrich K. H. Ecker|title=Neutralizing misinformation through inoculation: Exposing misleading argumentation techniques reduces their influence|journal=]|date=2017-05-05|volume= 12|issue= 5|pages= e0175799|doi=10.1371/journal.pone.0175799|pmid= 28475576|pmc= 5419564|bibcode= 2017PLoSO..1275799C|doi-access=free}}
* Helfand, David J., ''A Survival Guide to the Misinformation Age: Scientific Habits of Mind''. Columbia University Press, 2016. {{ISBN|978-0231541022}}
* Christopher Murphy (2005). Competitive Intelligence: Gathering, Analysing And Putting It to Work. Gower Publishing, Ltd.. pp.&nbsp;186–189. {{ISBN|0-566-08537-2}}. A case study of misinformation arising from simple error * Christopher Murphy (2005). Competitive Intelligence: Gathering, Analysing And Putting It to Work. Gower Publishing, Ltd.. pp.&nbsp;186–189. {{ISBN|0-566-08537-2}}. A case study of misinformation arising from simple error
* O'Connor, Cailin, and James Owen Weatherall, "Why We Trust Lies: The most effective misinformation starts with seeds of ]", '']'', vol. 321, no. 3 (September 2019), pp.&nbsp;54–61. * {{cite news |last1=O'Connor |first1=Cailin |last2=Weatherall |first2=James Owen |title=How Misinformation Spreads—and Why We Trust It |url=https://www.scientificamerican.com/article/how-misinformation-spreads-and-why-we-trust-it/ |work=Scientific American |date=1 September 2019 }}
* O'Connor, Cailin, and James Owen Weatherall, ''The Misinformation Age; How False Beliefs Spread''. Yale University Press, 2019. {{ISBN|978-0300241006}}
* {{cite book |title=Bad Advice: Or Why Celebrities, Politicians, and Activists Aren't Your Best Source of Health Information |year=2019 |first=Paul |last=Offit |publisher=Columbia University Press |isbn=978-0-231-18699-5}}
* Persily, Nathaniel, and Joshua A. Tucker, eds. ''Social Media and Democracy: The State of the Field and Prospects for Reform''. Cambridge University Press, 2020. {{ISBN|978-1108858779}}
* {{cite journal |last1=Southwell |first1=Brian G. |last2=Brennen |first2=J. Scott Babwah |last3=Paquin |first3=Ryan |last4=Boudewyns |first4=Vanessa |last5=Zeng |first5=Jing |title=Defining and Measuring Scientific Misinformation |journal=] |date=2022 |volume=700 |issue=1 |pages=98–111 |doi=10.1177/00027162221084709 |doi-access=free}}
* Jürg Strässler (1982). Idioms in English: A Pragmatic Analysis. Gunter Narr Verlag. pp.&nbsp;43–44. {{ISBN|3-87808-971-6}}. * Jürg Strässler (1982). Idioms in English: A Pragmatic Analysis. Gunter Narr Verlag. pp.&nbsp;43–44. {{ISBN|3-87808-971-6}}.
* {{cite journal |last1=West |first1=Jevin D. |last2=Bergstrom |first2=Carl T. |title=Misinformation in and about science |journal=Proceedings of the National Academy of Sciences |date=2021 |volume=118 |issue=15 |doi=10.1073/pnas.1912444117 |pmid=33837146 |pmc=8054004 |bibcode=2021PNAS..11812444W |doi-access=free}}


==External links== == External links ==
{{Wikiquote}}
* (audio tutorial, graphic tutorial)
* {{cite web|url=https://www.npr.org/2020/04/17/837202898/comic-fake-news-can-be-deadly-heres-how-to-spot-it |title=Comic: Fake News Can Be Deadly. Here's How To Spot It |format=audio tutorial, graphic tutorial |website=NPR |
date=April 20, 2020 |
author1= Connie Hanzhang Jin | author2= Miles Parks }}
* {{cite web|url=https://www.msicertified.com/free-training/misinformation-and-disinformation-training/ |website=Management and Strategy Institute|title= Free Misinformation and Disinformation Training online |date=23 August 2022 | format=free online class }}


{{Disinformation}} {{Disinformation}}
{{Authority control}}


] ]
]
] ]
] ]

Latest revision as of 21:07, 31 December 2024

Incorrect or misleading information Not to be confused with Disinformation or Misinformation effect.
A sign campaigning for the successful Vote Leave in the 2016 United Kingdom European Union membership referendum. The claim made by the sign was widely considered to have been an example of misinformation.

Misinformation is incorrect or misleading information. Misinformation and disinformation are not interchangeable terms: Misinformation can exist with or without specific malicious intent whereas disinformation is distinct in that the information is deliberately deceptive and propagated. Misinformation can include inaccurate, incomplete, misleading, or false information as well as selective or half-truths. In January 2024, the World Economic Forum identified misinformation and disinformation, propagated by both internal and external interests, to "widen societal and political divides" as the most severe global risks within the next two years.

Much research on how to correct misinformation has focused on fact-checking. However, this can be challenging because the information deficit model does not necessarily apply well to beliefs in misinformation. Various researchers have also investigated what makes people susceptible to misinformation. People may be more prone to believe misinformation because they are emotionally connected to what they are listening to or are reading. Social media has made information readily available to society at anytime, and it connects vast groups of people along with their information at one time. Advances in technology have impacted the way people communicate information and the way misinformation is spread. Misinformation can influence people's beliefs about communities, politics, medicine, and more. The term also has the potential to be used to obfuscate legitimate speech and warp political discourses.

The term came into wider recognition during the mid-1990s through the early 2020s, when its effects on public ideological influence began to be investigated. However, misinformation campaigns have existed for hundreds of years.

Terminology

Misinformation is often used as an umbrella term to refer to many types of false information; more specifically it may refer to false information that is not shared to intentionally deceive or cause harm. Those who do not know that a piece of information is untrue, for instance, might disseminate it on social media in an effort to help.

Disinformation is created or spread by a person or organization actively attempting to deceive their audience. In addition to causing harm directly, disinformation can also cause indirect harm by undermining trust and obstructing the capacity to effectively communicate information with one another. Disinformation might consist of information that is partially or completely fabricated, taken out of context on purpose, exaggerated, or omits crucial details. Disinformation can appear in any medium including text, audio, and imagery. The distinction between mis- and dis-information can be muddy because the intent of someone sharing false information can be difficult to discern.

Malinformation is accurate information that is disseminated with malicious intent. This includes sensitive material that is disseminated in order to hurt someone or their reputation. Examples include doxing, revenge porn, and editing videos to remove important context or content.

Misinformation is information that was originally thought to be true but was later discovered not to be true, and often applies to emerging situations in which there is a lack of verifiable information or changing scientific understanding. For example, the scientific guidance around infant sleep positions has evolved over time, and these changes could be a source of confusion for new parents. Misinformation can also often be observed as news events are unfolding and questionable or unverified information fills information gaps. Even if later retracted, false information can continue to influence actions and memory.

Rumors are unverified information not attributed to any particular source and may be either true or false.

Definitions of these terms may vary between cultural contexts.

History

Early examples include the insults and smears spread among political rivals in Imperial and Renaissance Italy in the form of pasquinades. These are anonymous and witty verses named for the Pasquino piazza and talking statues in Rome. In pre-revolutionary France, "canards", or printed broadsides, sometimes included an engraving to convince readers to take them seriously.

During the summer of 1587, continental Europe anxiously awaited news as the Spanish Armada sailed to fight the English. The Spanish postmaster and Spanish agents in Rome promoted reports of Spanish victory in hopes of convincing Pope Sixtus V to release his promised one million ducats upon landing of troops. In France, the Spanish and English ambassadors promoted contradictory narratives in the press, and a Spanish victory was incorrectly celebrated in Paris, Prague, and Venice. It was not until late August that reliable reports of the Spanish defeat arrived in major cities and were widely believed; the remains of the fleet returned home in the autumn.

A lithograph from the first large scale spread of disinformation in America, the Great Moon Hoax

The first recorded large-scale disinformation campaign was the Great Moon Hoax, published in 1835 in the New York The Sun, in which a series of articles claimed to describe life on the Moon, "complete with illustrations of humanoid bat-creatures and bearded blue unicorns". The challenges of mass-producing news on a short deadline can lead to factual errors and mistakes. An example of such is the Chicago Tribune's infamous 1948 headline "Dewey Defeats Truman".

Social media platforms allow for easy spread of misinformation. Post-election surveys in 2016 suggest that many individuals who intake false information on social media believe them to be factual. The specific reasons why misinformation spreads through social media so easily remain unknown. A 2018 study of Twitter determined that, compared to accurate information, false information spread significantly faster, further, deeper, and more broadly. Similarly, a research study of Facebook found that misinformation was more likely to be clicked on than factual information.

Harry S. Truman displaying the inaccurate Chicago Tribune headline, an example of misinformation

Moreover, the advent of the Internet has changed traditional ways that misinformation spreads. During the 2016 United States presidential election, content from websites deemed 'untrustworthy' reached up to 40% of Americans, despite misinformation making up only 6% of overall news media. Misinformation has been spread during many health crises. For example, misinformation about alternative treatments was spread during the Ebola outbreak in 2014–2016. During the COVID-19 pandemic, the proliferation of mis- and dis-information was exacerbated by a general lack of health literacy.

Causes

Factors that contribute to beliefs in misinformation are an ongoing subject of study. According to Scheufele and Krause, misinformation belief has roots at the individual, group and societal levels. At the individual level, individuals have varying levels of skill in recognizing mis- or dis-information and may be predisposed to certain misinformation beliefs due to other personal beliefs, motivations, or emotions. However, evidence for the hypotheses that believers in misinformation use more cognitive heuristics and less effortful processing of information have produced mixed results. At the group level, in-group bias and a tendency to associate with like-minded or similar people can produce echo chambers and information silos that can create and reinforce misinformation beliefs. At the societal level, public figures like politicians and celebrities can disproportionately influence public opinions, as can mass media outlets. In addition, societal trends like political polarization, economic inequalities, declining trust in science, and changing perceptions of authority contribute to the impact of misinformation.

Historically, people have relied on journalists and other information professionals to relay facts. As the number and variety of information sources has increased, it has become more challenging for the general public to assess their credibility. This growth of consumer choice when it comes to news media allows the consumer to choose a news source that may align with their biases, which consequently increases the likelihood that they are misinformed. 47% of Americans reported social media as their main news source in 2017 as opposed to traditional news sources. Polling shows that Americans trust mass media at record-low rates, and that US young adults place similar levels of trust in information from social media and from national news organizations. The pace of the 24 hour news cycle does not always allow for adequate fact-checking, potentially leading to the spread of misinformation. Further, the distinction between opinion and reporting can be unclear to viewers or readers.

Sources of misinformation can appear highly convincing and similar to trusted legitimate sources. For example, misinformation cited with hyperlinks has been found to increase readers' trust. Trust is even higher when these hyperlinks are to scientific journals, and higher still when readers do not click on the sources to investigate for themselves. Research has also shown that the presence of relevant images alongside incorrect statements increases both their believability and shareability, even if the images do not actually provide evidence for the statements. For example, a false statement about macadamia nuts accompanied by an image of a bowl of macadamia nuts tends to be rated as more believable than the same statement without an image.

The translation of scientific research into popular reporting can also lead to confusion if it flattens nuance, sensationalizes the findings, or places too much emphasis on weaker levels of evidence. For instance, researchers have found that newspapers are more likely than scientific journals to cover observational studies and studies with weaker methodologies. Dramatic headlines may gain readers' attention, but they do not always accurately reflect scientific findings.

Human cognitive tendencies can also be a contributing factor to misinformation belief. One study found that an individual's recollection of political events could be altered when presented with misinformation about the event, even when primed to identify warning signs of misinformation. Misinformation may also be appealing by seeming novel or incorporating existing stereotypes.

Identification

Research has yielded a number of strategies that can be employed to identify misinformation, many of which share common features. According to Anne Mintz, editor of Web of Deception: Misinformation on the Internet, one of the simplest ways to determine whether information is factual is to use common sense. Mintz advises that the reader check whether the information makes sense and whether the source or sharers of the information might be biased or have an agenda. However, because emotions and preconceptions heavily impact belief, this is not always a reliable strategy. Readers tend to distinguish between unintentional misinformation and uncertain evidence from politically or financially motivated misinformation. The perception of misinformation depends on the political spectrum, with right-wing readers more concerned with attempts to hide reality. It can be difficult to undo the effects of misinformation once individuals believe it to be true. Individuals may desire to reach a certain conclusion, causing them to accept information that supports that conclusion, and are more likely to retain and share information if it emotionally resonates with them.

The SIFT Method, also called the Four Moves, is one commonly taught method of distinguishing between reliable and unreliable information. This method instructs readers to first Stop and begin to ask themselves about what they are reading or viewing - do they know the source and if it is reliable? Second, readers should Investigate the source. What is the source's relevant expertise and do they have an agenda? Third, a reader should Find better coverage and look for reliable coverage on the claim at hand to understand if there is a consensus around the issue. Finally, a reader should Trace claims, quotes, or media to their original context: has important information been omitted, or is the original source questionable?

Visual misinformation presents particular challenges, but there are some effective strategies for identification. Misleading graphs and charts can be identified through careful examination of the data presentation; for example, truncated axes or poor color choices can cause confusion. Reverse image searching can reveal whether images have been taken out of their original context. There are currently some somewhat reliable ways to identify AI-generated imagery, but it is likely that this will become more difficult to identify as the technology advances.

A person's formal education level and media literacy do correlate with their ability to recognize misinformation. People who are familiar with a topic, the processes of researching and presenting information, or have critical evaluation skills are more likely to correctly identify misinformation. However, these are not always direct relationships. Higher overall literacy does not always lead to improved ability to detect misinformation. Context clues can also significantly impact people's ability to detect misinformation.

Martin Libicki, author of Conquest In Cyberspace: National Security and Information Warfare, notes that readers should aim to be skeptical but not cynical. Readers should not be gullible, believing everything they read without question, but also should not be paranoid that everything they see or read is false.

Countermeasures

See also: Conspiracy theories § Interventions

Factors that contribute to the effectiveness of a corrective message include an individual's mental model or worldview, repeated exposure to the misinformation, time between misinformation and correction, credibility of the sources, and relative coherency of the misinformation and corrective message. Corrective messages will be more effective when they are coherent and/or consistent with the audience's worldview. They will be less effective when misinformation is believed to come from a credible source, is repeated prior to correction (even if the repetition occurs in the process of debunking), and/or when there is a time lag between the misinformation exposure and corrective message. Additionally, corrective messages delivered by the original source of the misinformation tend to be more effective. However, misinformation research has often been criticized for its emphasis on efficacy (i.e., demonstrating effects of interventions in controlled experiments) over effectiveness (i.e., confirming real-world impacts of these interventions). Critics argue that while laboratory settings may show promising results, these do not always translate into practical, everyday situations where misinformation spreads. Research has identified several major challenges in this field: an overabundance of lab research and a lack of field studies, the presence of testing effects that impede intervention longevity and scalability, modest effects for small fractions of relevant audiences, reliance on item evaluation tasks as primary efficacy measures, low replicability in the Global South and a lack of audience-tailored interventions, and the underappreciation of potential unintended consequences of intervention implementation.

Fact-checking and debunking

Websites have been created to help people to discern fact from fiction. For example, the site FactCheck.org aims to fact check the media, especially viral political stories. The site also includes a forum where people can openly ask questions about the information. Similar sites allow individuals to copy and paste misinformation into a search engine and the site will investigate it. Some sites exist to address misinformation about specific topics, such as climate change misinformation. DeSmog, formerly The DeSmogBlog, publishes factually accurate information in order to counter the well-funded disinformation campaigns spread by motivated deniers of climate change. Science Feedback focuses on evaluating science, health, climate, and energy claims in the media and providing an evidence-based analysis of their veracity.

Flagging or eliminating false statements in media using algorithmic fact checkers is becoming an increasingly common tactic to fight misinformation. Google and many social media platforms have added automatic fact-checking programs to their sites and created the option for users to flag information that they think is false. Google provides supplemental information pointing to fact-checking websites in search results for controversial topics. On Facebook, algorithms may warn users if what they are about to share is likely false. In some cases social media platforms' efforts to curb the spread of misinformation has resulted in controversy, drawing criticism from people who see these efforts as constructing a barrier to their right to expression.

One-on-one correction

Within the context of personal interactions, some strategies for debunking have the potential to be effective. Simply delivering facts is frequently ineffective because misinformation belief is often not the result of a deficit of accurate information, although individuals may be more likely to change their beliefs in response to information shared by someone with whom they have close social ties, like a friend or family member. More effective strategies focus on instilling doubt and encouraging people to examine the roots of their beliefs. In these situations, tone can also play a role: expressing empathy and understanding can keep communication channels open. It is important to remember that beliefs are driven not just by facts but by emotion, worldview, intuition, social pressure, and many other factors.

Social correction

Fact-checking and debunking can be done in one-on-one interactions, but when this occurs on social media it is likely that other people may encounter and read the interaction, potentially learning new information from it or examining their own beliefs. This type of correction has been termed social correction. Researchers have identified three ways to increase the efficacy of these social corrections for observers. First, corrections should include a link to a credible source of relevant information, like an expert organization. Second, the correct information should be repeated, for example at the beginning and end of the comment or response. Third, an alternative explanation should be offered. An effective social correction in response to a statement that chili peppers can cure COVID-19 might look something like: “Hot peppers in your food, though very tasty, cannot prevent or cure COVID-19. The best way to protect yourself against the new coronavirus is to keep at least 1 meter away from others and to wash your hands frequently and thoroughly. Adding peppers to your soup won't prevent or cure COVID-19. Learn more from the WHO." Interestingly, while the tone of the correction may impact how the target of the correction receives the message and can increase engagement with a message, it is less likely to affect how others seeing the correction perceive its accuracy.

While social correction has the potential to reach a wider audience with correct information, it can also potentially amplify an original post containing misinformation.

Prebunking

Unfortunately, misinformation typically spreads more readily than fact-checking. Further, even if misinformation is corrected, that does not mean it is forgotten or does not influence people's thoughts. Another approach, called prebunking, aims to "inoculate" against misinformation by showing people examples of misinformation and how it works before they encounter it. While prebunking can involve fact-based correction, it focuses more on identifying common logical fallacies (e.g., emotional appeals to manipulate individuals' perceptions and judgments, false dichotomies, or ad hominem fallacies) and tactics used to spread misinformation as well as common misinformation sources. Research about the efficacy of prebunking has shown promising results.

Other interventions

A report by the Royal Society in the UK lists additional potential or proposed countermeasures:

  • Automated detection systems (e.g. to flag or add context and resources to content)
  • Provenance enhancing technology (i.e. better enabling people to determine the veracity of a claim, image, or video)
  • APIs for research (i.e. for usage to detect, understand, and counter misinformation)
  • Active bystanders (e.g. corrective commenting)
  • Community moderation (usually of unpaid and untrained, often independent, volunteers)
  • Anti-virals (e.g. limiting the number of times a message can be forwarded in privacy-respecting encrypted chats)
  • Collective intelligence (examples being Misplaced Pages where multiple editors refine encyclopedic articles, and question-and-answer sites where outputs are also evaluated by others similar to peer-review)
  • Trustworthy institutions and data
  • Media literacy (increasing citizens' ability to use ICTs to find, evaluate, create, and communicate information, an essential skill for citizens of all ages)
    • Media literacy is taught in Estonian public schools – from kindergarten through to high school – since 2010 and "accepted 'as important as writing or reading'"
    • New Jersey mandated K-12 students to learn information literacy
    • "Inoculation" via educational videos shown to adults is being explored

Broadly described, the report recommends building resilience to scientific misinformation and a healthy online information environment and not having offending content removed. It cautions that censorship could e.g. drive misinformation and associated communities "to harder-to-address corners of the internet".

Online misinformation about climate change can be counteracted through different measures at different stages. Prior to misinformation exposure, education and "inoculation" are proposed. Technological solutions, such as early detection of bots and ranking and selection algorithms are suggested as ongoing mechanisms. Post misinformation, corrective and collaborator messaging can be used to counter climate change misinformation. Incorporating fines and similar consequences has also been suggested.

The International Panel on the Information Environment was launched in 2023 as a consortium of over 250 scientists working to develop effective countermeasures to misinformation and other problems created by perverse incentives in organizations disseminating information via the Internet.

There also is research and development of platform-built-in as well browser-integrated (currently in the form of addons) misinformation mitigation. This includes quality/neutrality/reliability ratings for news sources. Misplaced Pages's perennial sources page categorizes many large news sources by reliability. Researchers have also demonstrated the feasibility of falsity scores for popular and official figures by developing such for over 800 contemporary elites on Twitter as well as associated exposure scores.

Strategies that may be more effective for lasting correction of false beliefs include focusing on intermediaries (such as convincing activists or politicians who are credible to the people who hold false beliefs, or promoting intermediaries who have the same identities or worldviews as the intended audience), minimizing the association of misinformation with political or group identities (such as providing corrections from nonpartisan experts, or avoiding false balance based on partisanship in news coverage), and emphasizing corrections that are hard for people to avoid or deny (such as providing information that the economy is unusually strong or weak, or describing the increased occurrence of extreme weather events in response to climate change denial).

Limitations

Main article: Belief perseverance

Interventions need to account for the possibility that misinformation can persist in the population even after corrections are published. Possible reasons include difficulty in reaching the right people and corrections not having long-term effects. For example, if corrective information is only published in science-focused publications and fact-checking websites, it may not reach the people who believe in misinformation since they are less likely to read those sources. In addition, successful corrections may not be persistent, particularly if people are re-exposed to misinformation at a later date.

It has been suggested that directly countering misinformation can be counterproductive, which is referred to as a "backfire effect", but in practice this is very rare. A 2020 review of the scientific literature on backfire effects found that there have been widespread failures to replicate their existence, even under conditions that would be theoretically favorable to observing them. Due to the lack of reproducibility, as of 2020 most researchers believe that backfire effects are either unlikely to occur on the broader population level, or they only occur in very specific circumstances, or they do not exist. Brendan Nyhan, one of the researchers who initially proposed the occurrence of backfire effects, wrote in 2021 that the persistence of misinformation is most likely due to other factors. For most people, corrections and fact-checking are very unlikely to have a negative impact, and there is no specific group of people in which backfire effects have been consistently observed. In many cases, when backfire effects have been discussed by the media or by bloggers, they have been overgeneralized from studies on specific subgroups to incorrectly conclude that backfire effects apply to the entire population and to all attempts at correction.

There is an ongoing debate on whether misinformation interventions may have the negative side effect of reducing belief in both false and true information, regardless of veracity. For instance, one study found that inoculation and accuracy primes to some extent undermined users' ability to distinguish implausible from plausible conspiracy theories. Other scholars have shown through simulations that even if interventions reduce both the belief in false and true information, the effect on the media ecosystem may still be favorable due to different base rates in both beliefs.

Online misinformation

Further information: Health information on the Internet § Social media

In recent years, the proliferation of misinformation online has drawn widespread attention. More than half of the world's population had access to the Internet in the beginning of 2018. Digital and social media can contribute to the spread of misinformation – for instance, when users share information without first checking the legitimacy of the information they have found. People are more likely to encounter online information based on personalized algorithms. Google, Facebook and Yahoo News all generate newsfeeds based on the information they know about our devices, our location, and our online interests.

Although two people can search for the same thing at the same time, they are very likely to get different results based on what that platform deems relevant to their interests, fact or false. Various social media platforms have recently been criticized for encouraging the spread of false information, such as hoaxes, false news, and mistruths. It is responsible with influencing people's attitudes and judgment during significant events by disseminating widely believed misinformation. Furthermore, online misinformation can occur in numerous ways, including rumors, urban legends, factoids, etc. However, the underlying factor is that it contains misleading or inaccurate information.

Moreover, users of social media platforms may experience intensely negative feelings, perplexity, and worry as a result of the spread of false information. According to a recent study, one in ten Americans has gone through mental or emotional stress as a result of misleading information posted online. Spreading false information can also seriously impede the effective and efficient use of the information available on social media. An emerging trend in the online information environment is "a shift away from public discourse to private, more ephemeral, messaging", which is a challenge to counter misinformation.

On social media

Pew Research reports shared that approximately one in four American adults admitted to sharing misinformation on their social media platforms.

In the Information Age, social networking sites have become a notable agent for the spread of misinformation, fake news, and propaganda. Social media sites have changed their algorithms to prevent the spread of fake news but the problem still exists.

Image posts are the biggest spread of misinformation on social media, a fact which is grossly unrepresented in research. This leads to a "yawning gap of knowledge" as there is a collective ignorance on how harmful image-based posts are compared to other types of misinformation.

Spread

Social media platforms allow for easy spread of misinformation. The specific reasons why misinformation spreads through social media so easily remain unknown.

Agent-based models and other computational models have been used by researchers to explain how false beliefs spread through networks. Epistemic network analysis is one example of a computational method for evaluating connections in data shared in a social media network or similar network.

Researchers fear that misinformation in social media is "becoming unstoppable". It has also been observed that misinformation and disinformation reappear on social media sites.

Misinformation spread by bots has been difficult for social media platforms to address. Sites such as Facebook have algorithms that have been proven to further the spread of misinformation in which how content is spread among subgroups.

Social causes and echo chambers

Spontaneous spread of misinformation on social media usually occurs from users sharing posts from friends or mutually-followed pages. These posts are often shared from someone the sharer believes they can trust. Misinformation introduced through a social format influences individuals drastically more than misinformation delivered non-socially.

People are inclined to follow or support like-minded individuals, creating echo chambers and filter bubbles. Untruths or general agreement within isolated social clusters are difficult to counter. Some argue this causes an absence of a collective reality. Research has also shown that viral misinformation may spread more widely as a result of echo chambers, as the echo chambers provide an initial seed which can fuel broader viral diffusion.

Misinformation might be created and spread with malicious intent for reasons such as causing anxiety or deceiving audiences. Rumors created with or without malicious intent may be unknowingly shared by users. People may know what the scientific community has proved as a fact, and still refuse to accept it as such.

Lack of regulation

Misinformation on social media spreads quickly in comparison to traditional media because of the lack of regulation and examination required before posting.

Social media sites provide users with the capability to spread information quickly to other users without requiring the permission of a gatekeeper such as an editor, who might otherwise require confirmation of the truth before allowing publication.

The problem of misinformation in social media is getting worse as younger generations prefer social media over journalistic for their source of information.

Countermeasures

Combating the spread of misinformation on social medias is difficult for reasons such as :

  • the profusion of misinformation sources makes the reader's task of weighing the reliability of information more challenging
  • social media's propensity for culture wars embeds misinformation with identity-based conflict
  • the proliferation of echo chambers form an epistemic environment in which participants encounter beliefs and opinions that coincide with their own, moving the entire group toward more extreme positions.

With the large audiences that can be reached and the experts on various subjects on social media, some believe social media could also be the key to correcting misinformation.

Journalists today are criticized for helping to spread false information on these social platforms, but research shows they also play a role in curbing it through debunking and denying false rumors.

COVID-19 misinformation

Main article: COVID-19 misinformation

During the COVID-19 Pandemic, social media was used as one of the main propagators for spreading misinformation about symptoms, treatments, and long-term health-related problems. This problem has initialized a significant effort in developing automated detection methods for misinformation on social media platforms.

The creator of the Stop Mandatory Vaccination made money posting anti-vax false news on social media. He posted more than 150 posts aimed towards women, garnering a total of 1.6 million views and earning money for every click and share.

Misinformation on TikTok

A research report by NewsGuard found there is a very high level (~20% in their probes of videos about relevant topics) of online misinformation delivered – to a mainly young user base – with TikTok, whose (essentially unregulated) usage is increasing as of 2022.

Misinformation on Facebook

Main article: Misinformation on Facebook

A research study of Facebook found that misinformation was more likely to be clicked on than factual information. The most common reasons that Facebook users were sharing misinformation for socially-motivated reasons, rather than taking the information seriously.

Facebook's coverage of misinformation has become a hot topic with the spread of COVID-19, as some reports indicated Facebook recommended pages containing health misinformation. For example, this can be seen when a user likes an anti-vax Facebook page. Automatically, more and more anti-vax pages are recommended to the user. Additionally, some reference Facebook's inconsistent censorship of misinformation leading to deaths from COVID-19.

Facebook estimated the existence of up to 60 million troll bots actively spreading misinformation on their platform, and has taken measures to stop the spread of misinformation, resulting in a decrease, though misinformation continues to exist on the platform. On Facebook, adults older than 65 were seven times more likely to share fake news than adults ages 18–29.

Misinformation on Twitter

Main articles: Twitter under Elon Musk § Misinformation and disinformation, and Community Notes

Twitter is one of the most concentrated platforms for engagement with political fake news. 80% of fake news sources are shared by 0.1% of users, who are "super-sharers". Older, more conservative social users are also more likely to interact with fake news. Another source of misinformation on Twitter are bot accounts, especially surrounding climate change. Bot accounts on Twitter accelerate true and fake news at the same rate. A 2018 study of Twitter determined that, compared to accurate information, false information spread significantly faster, further, deeper, and more broadly. A research study watched the process of thirteen rumors appearing on Twitter and noticed that eleven of those same stories resurfaced multiple times, after time had passed.

A social media app called Parler has caused much chaos as well. Right winged Twitter users who were banned on the app moved to Parler after the January 6 United States Capitol attack, and the app was being used to plan and facilitate more illegal and dangerous activities. Google and Apple later pulled the app off their respective app stores. This app has been able to cause a lot of misinformation and bias in the media, allowing for more political mishaps.

Misinformation on YouTube

This section is an excerpt from Social impact of YouTube § False scientific content.

Anti-intellectual beliefs flourish on YouTube. One well-publicized example is the network of content creators supporting the view that the Earth is flat, not a sphere. Researchers found that the YouTubers publishing "Flat Earth" content aim to polarize their audiences through arguments that build upon an anti-scientific narrative.

A study published in July 2019 concluded that most climate change-related videos support worldviews that are opposed to the scientific consensus on climate change. Though YouTube claimed in December 2019 that new recommendation policies reduced "borderline" recommendations by 70%, a January 2020 Avaaz study found that, for videos retrieved by the search terms "climate change", "global warming", and "climate manipulation", YouTube's "up next" sidebar presented videos containing information contradicting the scientific consensus 8%, 16% and 21% of the time, respectively. Avaaz argued that this "misinformation rabbit hole" means YouTube helps to spread climate denialism, and profits from it.

In November 2020, YouTube issued a one-week suspension of the account of One America News Network and permanently de-monetized its videos because of OANN's repeated violations of YouTube's policy prohibiting videos claiming sham cures for COVID-19. Without evidence, OANN also cast doubt on the validity of the 2020 U.S. presidential election.

On August 1, 2021, YouTube barred Sky News Australia from uploading new content for a week for breaking YouTube's rules on spreading COVID-19 misinformation. In September 2021, more than a year after YouTube said it would take down misinformation about the coronavirus vaccines, the accounts of six out of twelve anti-vaccine activists identified by the nonprofit Center for Countering Digital Hate were still searchable and still posting videos.

In October 2021, YouTube's owner Google announced it would no longer permit YouTube creators to earn advertising money for content that "contradicts well-established scientific consensus around the existence and causes of climate change", and that it will not allow ads that promote such views. In spite of this policy, many videos that included misinformation about climate change were not de-monetized. Earlier, climate change deniers' online YouTube content focused on denying global warming, or saying such warming isn't caused by humans burning fossil fuel. As such denials became untenable, using new tactics that evade YouTube's policies to combat misinformation, content shifted to asserting that climate solutions are not workable, saying global warming is harmless or even beneficial, and accusing the environmental movement of being unreliable.

Lack of peer review

Promoting more Peer Review to benefit the accuracy in information

Due to the decentralized nature and structure of the Internet, content creators can easily publish content without being required to undergo peer review, prove their qualifications, or provide backup documentation. While library books have generally been reviewed and edited by an editor, publishing company, etc., Internet sources cannot be assumed to be vetted by anyone other than their authors. Misinformation may be produced, reproduced, and posted immediately on most online platforms.

Censorship accusations

Social media sites such as Facebook and Twitter have found themselves defending accusations of censorship for removing posts they have deemed to be misinformation. Social media censorship policies relying on government agency-issued guidance to determine information validity have garnered criticism that such policies have the unintended effect of stifling dissent and criticism of government positions and policies. Most recently, social media companies have faced criticism over allegedly prematurely censoring the discussion of the SARS-CoV 2 Lab Leak Hypothesis.

Other accusations of censorship appear to stem from attempts to prevent social media consumers from self-harm through the use of unproven COVID-19 treatments. For example, in July 2020, a video went viral showing Dr. Stella Immanuel claiming hydroxychloroquine was an effective cure for COVID-19. In the video, Immanuel suggested that there was no need for masks, school closures, or any kind of economic shut down; attesting that her alleged cure was highly effective in treating those infected with the virus. The video was shared 600,000 times and received nearly 20 million views on Facebook before it was taken down for violating community guidelines on spreading misinformation. The video was also taken down on Twitter overnight, but not before former president Donald Trump shared it to his page, which was followed by over 85 million Twitter users. NIAID director Dr. Anthony Fauci and members of the World Health Organization (WHO) quickly discredited the video, citing larger-scale studies of hydroxychloroquine showing it is not an effective treatment of COVID-19, and the FDA cautioned against using it to treat COVID-19 patients following evidence of serious heart problems arising in patients who have taken the drug.

Another prominent example of misinformation removal criticized by some as an example of censorship was the New York Post's report on the Hunter Biden laptops approximately two weeks before the 2020 presidential election, which was used to promote the Biden–Ukraine conspiracy theory. Social media companies quickly removed this report, and the Post's Twitter account was temporarily suspended. Over 50 intelligence officials found the disclosure of emails allegedly belonging to Joe Biden's son had all the "classic earmarks of a Russian information operation". Later evidence emerged that at least some of the laptop's contents were authentic.

Noteworthy examples

An example of bad information from media sources that led to the spread of misinformation occurred in November 2005, when Chris Hansen on Dateline NBC claimed that law enforcement officials estimate 50,000 predators are online at any moment. Afterward, the U.S. attorney general at the time, Alberto Gonzales, repeated the claim. However, the number that Hansen used in his reporting had no backing. Hansen said he received the information from Dateline expert Ken Lanning, but Lanning admitted that he made up the number 50,000 because there was no solid data on the number. According to Lanning, he used 50,000 because it sounds like a real number, not too big and not too small, and referred to it as a "Goldilocks number". Reporter Carl Bialik says that the number 50,000 is used often in the media to estimate numbers when reporters are unsure of the exact data.

During the COVID-19 pandemic, a conspiracy theory that COVID-19 was linked to the 5G network gained significant traction worldwide after emerging on social media.

Misinformation was a major talking point during the 2016 U.S. presidential election with claims of social media sites allowing "fake news" to be spread.

Impact

The Liar's Dividend describes a situation in which individuals are so concerned about realistic misinformation (in particular, deepfakes) that they begin to mistrust real content, particularly if someone claims that it is false. For instance, a politician could benefit from claiming that a real video of them doing something embarrassing was actually AI-generated or altered, leading followers to mistrust something that was actually real. On a larger scale this problem can lead to erosion in the public's trust of generally reliable information sources.

Misinformation can affect all aspects of life. Allcott, Gentzkow, and Yu concur that the diffusion of misinformation through social media is a potential threat to democracy and broader society. The effects of misinformation can lead to decline of accuracy of information as well as event details. When eavesdropping on conversations, one can gather facts that may not always be true, or the receiver may hear the message incorrectly and spread the information to others. On the Internet, one can read content that is stated to be factual but that may not have been checked or may be erroneous. In the news, companies may emphasize the speed at which they receive and send information but may not always be correct in the facts. These developments contribute to the way misinformation may continue to complicate the public's understanding of issues and to serve as a source for belief and attitude formation.

In regards to politics, some view being a misinformed citizen as worse than being an uninformed citizen. Misinformed citizens can state their beliefs and opinions with confidence and thus affect elections and policies. This type of misinformation occurs when a speaker appears "authoritative and legitimate", while also spreading misinformation. When information is presented as vague, ambiguous, sarcastic, or partial, receivers are forced to piece the information together and make assumptions about what is correct. Misinformation has the power to sway public elections and referendums if it gains enough momentum. Leading up to the 2016 UK European Union membership referendum, for example, a figure used prominently by the Vote Leave campaign claimed that by leaving the EU the UK would save £350 million a week, 'for the NHS'. Claims then circulated widely in the campaign that this amount would (rather than could theoretically) be redistributed to the British National Health Service after Brexit. This was later deemed a "clear misuse of official statistics" by the UK statistics authority.

Moreover, the advert infamously shown on the side of London's double-decker busses did not take into account the UK's budget rebate, and the idea that 100% of the money saved would go to the NHS was unrealistic. A poll published in 2016 by Ipsos MORI found that nearly half of the British public believed this misinformation to be true. Even when information is proven to be misinformation, it may continue to shape attitudes towards a given topic, meaning it has the power to swing political decisions if it gains enough traction. A study conducted by Soroush Vosoughi, Deb Roy and Sinan Aral looked at Twitter data including 126,000 posts spread by 3 million people over 4.5 million times. They found that political news traveled faster than any other type of information. They found that false news about politics reached more than 20,000 people three times faster than all other types of false news.

Aside from political propaganda, misinformation can also be employed in industrial propaganda. Using tools such as advertising, a company can undermine reliable evidence or influence belief through a concerted misinformation campaign. For instance, tobacco companies employed misinformation in the second half of the twentieth century to diminish the reliability of studies that demonstrated the link between smoking and lung cancer.

In the medical field, misinformation can immediately lead to life endangerment as seen in the case of the public's negative perception towards vaccines or the use of herbs instead of medicines to treat diseases. In regards to the COVID-19 pandemic, the spread of misinformation has proven to cause confusion as well as negative emotions such as anxiety and fear. Misinformation regarding proper safety measures for the prevention of the virus that go against information from legitimate institutions like the World Health Organization can also lead to inadequate protection and possibly place individuals at risk for exposure.

Some scholars and activists are heading movements to eliminate the mis/disinformation and information pollution in the digital world. One theory, "information environmentalism," has become a curriculum in some universities and colleges. The general study of misinformation and disinformation is by now also common across various academic disciplines, including sociology, communication, computer science, and political science, leading to the emerging field being described loosely as "Misinformation and Disinformation Studies". However, various scholars and journalists have criticised this development, pointing to problematic normative assumptions, a varying quality of output and lack of methodological rigor, as well as a too strong impact of mis- and disinformation research in shaping public opinion and policymaking. Summarising the most frequent points of critique, communication scholars Chico Camargo and Felix Simon wrote in an article for the Harvard Kennedy School Misinformation Review that "mis-/disinformation studies has been accused of lacking clear definitions, having a simplified understanding of what it studies, a too great emphasis on media effects, a neglect of intersectional factors, an outsized influence of funding bodies and policymakers on the research agenda of the field, and an outsized impact of the field on policy and policymaking."

AI contribution to the problem and aid in combatting

Artificial intelligence exacerbates the problem of misinformation but also contributes to the fight against misinformation.

AI's contribution to the problem

  • Deepfakes and Synthetic media create very convincing visual, audio, and textual evidence that is difficult to distinguish from legitimate authoritative evidence.
  • Internet bots and automated Internet trolls can rapidly sow disinformation.
  • Algorithmic bias plays a role in amplification of sensational and controversial material regardless of truth.

AI as a tool to combat misinformation

See also

References

  1. Henley, Jon (10 June 2016). "Why Vote Leave's £350m weekly EU cost claim is wrong". The Guardian. Retrieved 14 October 2024.
  2. "The UK's EU membership fee". Full Fact. 25 February 2016. Retrieved 14 October 2024.
  3. "Reality Check: Would Brexit mean extra £350m a week for NHS?". BBC News. 15 April 2016.
  4. Ackrill, Robert (27 April 2016). "Fact Check: how much does the UK actually pay to the EU?". The Conversation.
  5. ^ Merriam-Webster Dictionary (19 August 2020). "Misinformation". Archived from the original on 25 February 2019. Retrieved 19 August 2020.
  6. Fetzer, James H. (2004-05-01). "Information: Does it Have To Be True?". Minds and Machines. 14 (2): 223–229. doi:10.1023/B:MIND.0000021682.61365.56. ISSN 1572-8641. S2CID 31906034.
  7. Woolley, Samuel C.; Howard, Philip N. (2016). "Political Communication, Computational Propaganda, and Autonomous Agents". International Journal of Communication. 10: 4882–4890. Archived from the original on 2019-10-22. Retrieved 2019-10-22.
  8. ^ Caramancion, Kevin Matthe (2020). "An Exploration of Disinformation as a Cybersecurity Threat". 2020 3rd International Conference on Information and Computer Technologies (ICICT). pp. 440–444. doi:10.1109/icict50521.2020.00076. ISBN 978-1-7281-7283-5. S2CID 218651389.
  9. Karlova, Natascha A.; Fisher, Karen E. (2013-03-15). "A social diffusion model of misinformation and disinformation for understanding human information behaviour". informationr.net. Archived from the original on 2023-05-11. Retrieved 2023-05-11.
  10. ^ Fallis, Don (2015). "What Is Disinformation?". Library Trends. 63 (3): 401–426. doi:10.1353/lib.2015.0014. hdl:2142/89818. ISSN 1559-0682. S2CID 13178809.
  11. Diaz Ruiz, Carlos (2023-10-30). "Disinformation on digital media platforms: A market-shaping approach". New Media & Society. doi:10.1177/14614448231207644. ISSN 1461-4448.
  12. The Global Risks Report 2024, World Economic Forum. ISBN 978-2-940631-64-3
  13. ^ Lewandowsky, Stephan; Ecker, Ullrich K. H.; Seifert, Colleen M.; Schwarz, Norbert; Cook, John (2012). "Misinformation and Its Correction: Continued Influence and Successful Debiasing". Psychological Science in the Public Interest. 13 (3): 106–131. doi:10.1177/1529100612451018. JSTOR 23484653. PMID 26173286. S2CID 42633.
  14. ^ Diaz Ruiz, Carlos; Nilsson, Tomas (2022-08-08). "Disinformation and Echo Chambers: How Disinformation Circulates on Social Media Through Identity-Driven Controversies". Journal of Public Policy & Marketing. 42: 18–35. doi:10.1177/07439156221103852. ISSN 0743-9156.
  15. ^ Ecker, Ullrich K. H.; Lewandowsky, Stephan; Cook, John; Schmid, Philipp; Fazio, Lisa K.; Brashier, Nadia; Kendeou, Panayiota; Vraga, Emily K.; Amazeen, Michelle A. (January 2022). "The psychological drivers of misinformation belief and its resistance to correction". Nature Reviews Psychology. 1 (1): 13–29. doi:10.1038/s44159-021-00006-y. hdl:1983/889ddb0f-0d44-44f4-a54f-57c260ae4917. ISSN 2731-0574.
  16. ^ Aral 2020.
  17. ^ Blum, Dani (March 16, 2024). "Health Misinformation Is Evolving. Here's How to Spot It. - Experts offer tips for combating false medical claims in your own circles". The New York Times. Archived from the original on March 16, 2024. Retrieved March 17, 2024.
  18. Bode, Leticia; Vraga, Emily (23 June 2015). "In Related News, That was Wrong: The Correction of Misinformation Through Related Stories Functionality in Social Media". Journal of Communication. 65 (4): 619–638. doi:10.1111/jcom.12166. S2CID 142769329. Archived from the original on 2022-11-12. Retrieved 2024-01-31.
  19. Posetti, Julie; Matthews, Alice (June 23, 2018). "A Short Guide to the History of 'Fake News' and Disinformation: A New ICFJ Learning Module". International Center for Journalists. Archived from the original on 2019-02-25. Retrieved 2024-01-31.
  20. Greason, Grace (2018-07-06). "Information Disorder: The Definitional Toolbox". First Draft. Retrieved 2024-04-03.
  21. ^ Marquez, Natalie. "Research Guides: Misinformation – Get the Facts: What is Misinformation?". guides.lib.uci.edu. Archived from the original on 2023-03-16. Retrieved 2023-03-16.
  22. ^ "What is disinformation?". Die Bundesregierung informiert | Startseite (in German). Archived from the original on 2023-03-23. Retrieved 2023-03-16.
  23. "Foreign Influence Operations and Disinformation | Cybersecurity and Infrastructure Security Agency CISA". www.cisa.gov.
  24. Hanopol, Nica Rhiana (2021-05-12). "When 'misinformation' strikes, experts say here's what you can do". VERA Files. Retrieved 2024-04-03.
  25. "Safe to Sleep". Retrieved April 3, 2024.
  26. Ecker, Ullrich K.H.; Lewandowsky, Stephan; Cheung, Candy S.C.; Maybery, Murray T. (November 2015). "He did it! She did it! No, she did not! Multiple causal explanations and the continued influence of misinformation" (PDF). Journal of Memory and Language. 85: 101–115. doi:10.1016/j.jml.2015.09.002. Archived (PDF) from the original on 2022-12-07. Retrieved 2022-11-01.
  27. "Definition of RUMOR". www.merriam-webster.com. May 21, 2023. Archived from the original on February 4, 2022. Retrieved February 4, 2022.
  28. ^ Zeng, Jing; Chan, Chung-hong (2021-01-01). "A cross-national diagnosis of infodemics: comparing the topical and temporal features of misinformation around COVID-19 in China, India, the US, Germany and France". Online Information Review. 45 (4): 709–728. doi:10.1108/OIR-09-2020-0417. ISSN 1468-4527. S2CID 234242942. Archived from the original on 2024-02-14. Retrieved 2023-05-11.
  29. "The True History of Fake News". The New York Review of Books. 2017-02-13. Archived from the original on 2019-02-05. Retrieved 2019-02-24.
  30. Andrew Pettegree (2015). The Invention of News. Yale University Press. pp. 153–4. ISBN 978-0-300-21276-1.
  31. "A short guide to the history of 'fake news' and disinformation". International Center for Journalists. Archived from the original on 2019-02-25. Retrieved 2019-02-24.
  32. Jones, Tim (2020-10-31). "Dewey defeats Truman: The most famous wrong call in electoral history". Chicago Tribune. Retrieved 2024-04-03.
  33. Allcott, Hunt; Gentzkow, Matthew; Yu, Chuan (April 2019). "Trends in the diffusion of misinformation on social media". Research & Politics. 6 (2): 205316801984855. arXiv:1809.05901. doi:10.1177/2053168019848554. ISSN 2053-1680. S2CID 52291737.
  34. ^ Vosoughi, Soroush; Roy, Deb; Aral, Sinan (2018-03-09). "The spread of true and false news online". Science. 359 (6380): 1146–1151. Bibcode:2018Sci...359.1146V. doi:10.1126/science.aap9559. ISSN 0036-8075. PMID 29590045.
  35. Godfrey-Smith, Peter (December 1989). "Misinformation". Canadian Journal of Philosophy. 19 (4): 533–550. doi:10.1080/00455091.1989.10716781. ISSN 0045-5091. S2CID 246637810.
  36. West, Jevin D.; Bergstrom, Carl T. (2021-04-13). "Misinformation in and about science". Proceedings of the National Academy of Sciences. 118 (15): e1912444117. Bibcode:2021PNAS..11812444W. doi:10.1073/pnas.1912444117. ISSN 0027-8424. PMC 8054004. PMID 33837146.
  37. Fung, Isaac Chun-Hai; Fu, King-Wa; Chan, Chung-Hong; Chan, Benedict Shing Bun; Cheung, Chi-Ngai; Abraham, Thomas; Tse, Zion Tsz Ho (May 2016). "Social Media's Initial Reaction to Information and Misinformation on Ebola, August 2014: Facts and Rumors". Public Health Reports. 131 (3): 461–473. doi:10.1177/003335491613100312. ISSN 0033-3549. PMC 4869079. PMID 27252566.
  38. Oyeyemi, Sunday Oluwafemi; Gabarron, Elia; Wynn, Rolf (2014-10-14). "Ebola, Twitter, and misinformation: a dangerous combination?". BMJ. 349: g6178. doi:10.1136/bmj.g6178. ISSN 1756-1833. PMID 25315514. S2CID 731448. Archived from the original on 2022-11-01. Retrieved 2023-05-11.
  39. Swire-Thompson, Briony; Lazer, David (2020-04-02). "Public Health and Online Misinformation: Challenges and Recommendations". Annual Review of Public Health. 41 (1): 433–451. doi:10.1146/annurev-publhealth-040119-094127. ISSN 0163-7525. PMID 31874069.
  40. Jerit, Jennifer; Zhao, Yangzi (2020-05-11). "Political Misinformation". Annual Review of Political Science. 23 (1): 77–94. doi:10.1146/annurev-polisci-050718-032814. ISSN 1094-2939. S2CID 212733536.
  41. ^ Scheufele, Dietram; Krause, Nicole (April 16, 2019). "Science audiences, misinformation, and fake news". Proceedings of the National Academy of Sciences. 116 (16): 7662–7669. Bibcode:2019PNAS..116.7662S. doi:10.1073/pnas.1805871115. PMC 6475373. PMID 30642953.
  42. Pennycook, Gordon; Cheyne, James Allan; Seli, Paul; Koehler, Derek J.; Fugelsang, Jonathan A. (2012-06-01). "Analytic cognitive style predicts religious and paranormal belief". Cognition. 123 (3): 335–346. doi:10.1016/j.cognition.2012.03.003. ISSN 0010-0277. PMID 22481051.
  43. Robson, Samuel G.; Faasse, Kate; Gordon, Eliza-Rose; Jones, Samuel P.; Drew, Manisara; Martire, Kristy A. (2024-10-09). "Lazy or different? A quantitative content analysis of how believers and nonbelievers of misinformation reason". Advances.in/Psychology. 2: e003511. doi:10.56296/aip00027. ISSN 2976-937X.
  44. Scherer, Laura D.; McPhetres, Jon; Pennycook, Gordon; Kempe, Allison; Allen, Larry A.; Knoepke, Christopher E.; Tate, Channing E.; Matlock, Daniel D. (April 2021). "Who is susceptible to online health misinformation? A test of four psychosocial hypotheses". Health Psychology. 40 (4): 274–284. doi:10.1037/hea0000978. ISSN 1930-7810. PMID 33646806.
  45. Törnberg, Petter (2018-09-20). Bauch, Chris T. (ed.). "Echo chambers and viral misinformation: Modeling fake news as complex contagion". PLOS ONE. 13 (9): e0203958. Bibcode:2018PLoSO..1303958T. doi:10.1371/journal.pone.0203958. ISSN 1932-6203. PMC 6147442. PMID 30235239.
  46. "Literature Review: Echo chambers, filter bubbles and polarization" (PDF). Archived (PDF) from the original on 19 January 2022. Retrieved 21 February 2022.
  47. "Web of Deception: Misinformation on the Internet". The Electronic Library. 20 (6): 521. 1 December 2002. doi:10.1108/el.2002.20.6.521.7.
  48. "Students Have 'Dismaying' Inability To Tell Fake News From Real, Study Finds". NPR. Retrieved April 4, 2024.
  49. ^ Lewandowsky, Stephan; Ecker, Ullrich K. H.; Cook, John (December 2017). "Beyond misinformation: Understanding and coping with the 'post-truth' era" (PDF). Journal of Applied Research in Memory and Cognition. 6 (4): 353–369. doi:10.1016/j.jarmac.2017.07.008. hdl:1983/1b4da4f3-009d-4287-8e45-a0a1d7b688f7. S2CID 149003083. Archived (PDF) from the original on 2023-01-18. Retrieved 2022-11-01.
  50. Shearer, Elisa; Gottfried, Jeffrey (2017-09-07). "News Use Across Social Media Platforms 2017". Pew Research Center's Journalism Project. Archived from the original on 2021-03-16. Retrieved 2021-03-28.
  51. "Media Confidence in U.S. Matches 2016 Record Low". Gallup.com. 2023-10-19. Retrieved 2024-05-16.
  52. Liedke, Jacob; Gottfried, Jeffrey (2022-10-27). "U.S. adults under 30 now trust information from social media almost as much as from national news outlets". Pew Research Center. Retrieved 2024-05-16.
  53. Croteau, David; Hoynes, William; Milan, Stefania. "Media Technology" (PDF). Media Society: Industries, Images, and Audiences. pp. 285–321. Archived (PDF) from the original on January 2, 2013. Retrieved March 21, 2013.
  54. Mitchell, Amy; Gottfried, Jeffrey; Barthel, Michael; Sumida, Nami (2018-06-18). "Distinguishing Between Factual and Opinion Statements in the News". Pew Research Center. Retrieved 2024-05-16.
  55. "Journalists know news and opinion are separate, but readers often can't tell the difference". Nieman Lab. Retrieved 2024-05-16.
  56. "Stanford researchers find students have trouble judging the credibility of information online". Stanford Graduate School of Education. Stanford University. 2016-11-21. Retrieved 2024-04-03.
  57. Verma, Nitin; Fleischmann, Kenneth R.; Koltai, Kolina S. (January 2017). "Human values and trust in scientific journals, the mainstream media and fake news". Proceedings of the Association for Information Science and Technology. 54 (1): 426–435. doi:10.1002/pra2.2017.14505401046. S2CID 51958978.
  58. Sehgal, Vibhor; Peshin, Ankit; Afroz, Sadia; Farid, Hany (2021-04-20), Mutual Hyperlinking Among Misinformation Peddlers, arXiv:2104.11694
  59. ^ Newman, Eryn J.; Garry, Maryanne; Bernstein, Daniel M.; Kantner, Justin; Lindsay, D. Stephen (October 2012). "Nonprobative photographs (or words) inflate truthiness". Psychonomic Bulletin & Review. 19 (5): 969–974. doi:10.3758/s13423-012-0292-0. ISSN 1531-5320. PMID 22869334.
  60. Sundar, S Shyam; Molina, Maria D; Cho, Eugene (2021-11-12). "Seeing Is Believing: Is Video Modality More Powerful in Spreading Fake News via Online Messaging Apps?". Journal of Computer-Mediated Communication. 26 (6): 301–319. doi:10.1093/jcmc/zmab010. ISSN 1083-6101.
  61. Selvaraj, Senthil; Borkar, Durga S.; Prasad, Vinay (2014-01-17). "Media Coverage of Medical Journals: Do the Best Articles Make the News?". PLOS ONE. 9 (1): e85355. Bibcode:2014PLoSO...985355S. doi:10.1371/journal.pone.0085355. ISSN 1932-6203. PMC 3894978. PMID 24465543.
  62. "How to spot misleading science reporting – QB3 Berkeley". qb3.berkeley.edu. Retrieved 2024-04-12.
  63. Freeze, Melanie; Baumgartner, Mary; Bruno, Peter; Gunderson, Jacob R.; Olin, Joshua; Ross, Morgan Quinn; Szafran, Justine (2021-12-01). "Fake Claims of Fake News: Political Misinformation, Warnings, and the Tainted Truth Effect". Political Behavior. 43 (4): 1433–1465. doi:10.1007/s11109-020-09597-3. ISSN 1573-6687.
  64. Healy, Melissa (8 February 2020). "Misinformation about the coronavirus abounds, but correcting it can backfire". Los Angeles Times.
  65. Mintz, Anne. "The Misinformation Superhighway?". PBS. Archived from the original on 2 April 2013. Retrieved 26 February 2013.
  66. ^ Hameleers, Michael; Brosius, Anna (16 February 2022). "You Are Wrong Because I Am Right! The Perceived Causes and Ideological Biases of Misinformation Beliefs". International Journal of Public Opinion Research. 34 (1). doi:10.1093/ijpor/edab028. ISSN 0954-2892.
  67. Ecker, Ullrich K. H.; Lewandowsky, Stephan; Chadwick, Matthew (2020-04-22). "Can Corrections Spread Misinformation to New Audiences? Testing for the Elusive Familiarity Backfire Effect". Cognitive Research: Principles and Implications. 5 (1): 41. doi:10.31219/osf.io/et4p3. hdl:1983/0d5feec2-5878-4af6-b5c7-fbbd398dd4c4. PMC 7447737. PMID 32844338. Archived from the original on 2020-06-17. Retrieved 2020-06-17.
  68. Lewandowsky, Stephan; Ecker, Ullrich K. H.; Seifert, Colleen M.; Schwarz, Norbert; Cook, John (2012). "Misinformation and Its Correction: Continued Influence and Successful Debiasing". Psychological Science in the Public Interest. 13 (3): 106–131. doi:10.1177/1529100612451018. JSTOR 23484653. PMID 26173286. S2CID 42633.
  69. "SIFT (The Four Moves)". Hapgood. 2019-06-19. Retrieved 2024-04-12.
  70. Cowles, Kelsey; Miller, Rebekah; Suppok, Rachel (2024-01-02). "When Seeing Isn't Believing: Navigating Visual Health Misinformation through Library Instruction". Medical Reference Services Quarterly. 43 (1): 44–58. doi:10.1080/02763869.2024.2290963. ISSN 0276-3869. PMID 38237023.
  71. Grootendorst, Maarten (2021-03-23). "Misleading Graphs". Maarten Grootendorst. Retrieved 2024-04-12.
  72. Qian, Sijia; Shen, Cuihua; Zhang, Jingwen (2022-11-04). "Fighting cheapfakes: using a digital media literacy intervention to motivate reverse search of out-of-context visual misinformation". Journal of Computer-Mediated Communication. 28 (1). doi:10.1093/jcmc/zmac024. ISSN 1083-6101.
  73. Groh, Matthew; Epstein, Ziv; Firestone, Chaz; Picard, Rosalind (2022-01-05). "Deepfake detection by human crowds, machines, and machine-informed crowds". Proceedings of the National Academy of Sciences. 119 (1). arXiv:2105.06496. Bibcode:2022PNAS..11910013G. doi:10.1073/pnas.2110013119. ISSN 0027-8424. PMC 8740705. PMID 34969837.
  74. Bond, Shannon. "AI-generated images are everywhere. Here's how to spot them". NPR. Retrieved April 12, 2024.
  75. Hern, Alex (2024-04-08). "'Time is running out': can a future of undetectable deepfakes be avoided?". The Guardian. ISSN 0261-3077. Retrieved 2024-04-12.
  76. Verma, Pranshu. "AI Can Draw Hands Now". The Washington Post.
  77. Khan, M. Laeeq; Idris, Ika Karlina (2 December 2019). "Recognise misinformation and verify before sharing: a reasoned action and information literacy perspective". Behaviour & Information Technology. 38 (12): 1194–1212. doi:10.1080/0144929x.2019.1578828. S2CID 86681742.
  78. ^ Lazer, David M. J.; Baum, Matthew A.; Benkler, Yochai; Berinsky, Adam J.; Greenhill, Kelly M.; Menczer, Filippo; Metzger, Miriam J.; Nyhan, Brendan; Pennycook, Gordon; Rothschild, David; Schudson, Michael; Sloman, Steven A.; Sunstein, Cass R.; Thorson, Emily A.; Watts, Duncan J.; Zittrain, Jonathan L. (2018). "The science of fake news". Science. 359 (6380): 1094–1096. arXiv:2307.07903. Bibcode:2018Sci...359.1094L. doi:10.1126/science.aao2998. PMID 29590025. S2CID 4410672.
  79. Vraga, Emily K.; Bode, Leticia (December 2017). "Leveraging institutions, educators, and networks to correct misinformation: A commentary on Lewandosky, Ecker, and Cook". Journal of Applied Research in Memory and Cognition. 6 (4): 382–388. doi:10.1016/j.jarmac.2017.09.008.
  80. Caramancion, Kevin Matthe (2020). "Understanding the Impact of Contextual Clues in Misinformation Detection". 2020 IEEE International IOT, Electronics and Mechatronics Conference (IEMTRONICS). pp. 1–6. doi:10.1109/IEMTRONICS51293.2020.9216394. ISBN 978-1-7281-9615-2. S2CID 222297695.
  81. Libicki, Martin (2007). Conquest in Cyberspace: National Security and Information Warfare. New York: Cambridge University Press. pp. 51–55. ISBN 978-0-521-87160-0.
  82. Walter, Nathan; Tukachinsky, Riva (March 2020). "A Meta-Analytic Examination of the Continued Influence of Misinformation in the Face of Correction: How Powerful Is It, Why Does It Happen, and How to Stop It?". Communication Research. 47 (2): 155–177. doi:10.1177/0093650219854600. S2CID 197731687. Archived from the original on 2022-12-07. Retrieved 2022-11-01.
  83. ^ Roozenbeek, Jon; Remshard, Miriam; Kyrychenko, Yara (2024-07-27). "Beyond the headlines: On the efficacy and effectiveness of misinformation interventions". Advances.in/Psychology. 2: e24569. doi:10.56296/aip00019. ISSN 2976-937X.
  84. Tay, Li Qian; Lewandowsky, Stephan; Hurlstone, Mark J.; Kurz, Tim; Ecker, Ullrich K. H. (2023-10-05). "A focus shift in the evaluation of misinformation interventions". Harvard Kennedy School Misinformation Review. 4 (5). doi:10.37016/mr-2020-124. PMC 7616579. PMID 39301480.
  85. "Ask FactCheck". www.factcheck.org. Archived from the original on 2016-03-31. Retrieved 2016-03-31.
  86. ^ Fernandez, Miriam; Alani, Harith (2018). "Online Misinformation". Companion of the Web Conference 2018 on the Web Conference 2018 – WWW '18. ACM Press. pp. 595–602. doi:10.1145/3184558.3188730. ISBN 978-1-4503-5640-4. S2CID 13799324.
  87. "About - Science Feedback". science.feedback.org/. Retrieved 2024-04-12.
  88. Nott, Lata (2022-08-03). "Free Speech on Social Media: The Complete Guide". Freedom Forum. Retrieved 2024-04-12.
  89. Pasquetto, Irene V.; Jahani, Eaman; Atreja, Shubham; Baum, Matthew (2022-04-07). "Social Debunking of Misinformation on WhatsApp: The Case for Strong and In-group Ties". Proceedings of the ACM on Human-Computer Interaction. 6 (CSCW1): 117:1–117:35. doi:10.1145/3512964.
  90. "How should you talk to friends and relatives who believe conspiracy theories?". 2020-12-21. Retrieved 2024-04-22.
  91. ^ Vraga, Emily K.; Bode, Leticia (October 2020). "Correction as a Solution for Health Misinformation on Social Media". American Journal of Public Health. 110 (S3): S278 – S280. doi:10.2105/AJPH.2020.305916. ISSN 0090-0036. PMC 7532323. PMID 33001724.
  92. "COVID-19 Mythbusters – World Health Organization". www.who.int. Retrieved 2024-04-22.
  93. MacKay, Melissa; Cimino, Andrea; Yousefinaghani, Samira; McWhirter, Jennifer E.; Dara, Rozita; Papadopoulos, Andrew (2022-06-06). "Canadian COVID-19 Crisis Communication on Twitter: Mixed Methods Research Examining Tweets from Government, Politicians, and Public Health for Crisis Communication Guiding Principles and Tweet Engagement". International Journal of Environmental Research and Public Health. 19 (11): 6954. doi:10.3390/ijerph19116954. ISSN 1660-4601. PMC 9180105. PMID 35682537.
  94. Bode, Leticia; Vraga, Emily K.; Tully, Melissa (2020-06-11). "Do the right thing: Tone may not affect correction of misinformation on social media". Harvard Kennedy School Misinformation Review. doi:10.37016/mr-2020-026.
  95. "Health Misinformation — Current Priorities of the U.S. Surgeon General". www.hhs.gov. Retrieved 2024-04-22.
  96. Smith, Rory (2020-12-18). "The "broadcast" model no longer works in an era of disinformation". First Draft. Retrieved 2024-04-12.
  97. ^ Garcia, Laura (2021-06-29). "A guide to prebunking: a promising way to inoculate against misinformation". First Draft. Retrieved 2024-04-12.
  98. Cook, John; Ellerton, Peter; Kinkead, David (1 February 2018). "Deconstructing climate misinformation to identify reasoning errors". Environmental Research Letters. 13 (2): 024018. Bibcode:2018ERL....13b4018C. doi:10.1088/1748-9326/aaa49f. S2CID 149353744.
  99. Traberg, Cecilie; Morton, Thomas; van der Linden, Sander (2024-06-20). "Counteracting socially endorsed misinformation through an emotion-fallacy inoculation". Advances.in/Psychology. 2: e765332. doi:10.56296/aip00017. ISSN 2976-937X.
  100. Roozenbeek, Jon; van der Linden, Sander; Goldberg, Beth; Rathje, Steve; Lewandowsky, Stephan (2022-08-26). "Psychological inoculation improves resilience against misinformation on social media". Science Advances. 8 (34): eabo6254. Bibcode:2022SciA....8O6254R. doi:10.1126/sciadv.abo6254. ISSN 2375-2548. PMC 9401631. PMID 36001675.
  101. Roozenbeek, Jon; van der Linden, Sander (2019-06-25). "Fake news game confers psychological resistance against online misinformation". Palgrave Communications. 5 (1): 1–10. doi:10.1057/s41599-019-0279-9. ISSN 2055-1045.
  102. ^ The online information environment: Understanding how the internet shapes people's engagement with scientific information (PDF). The Royal Society. January 2022. ISBN 978-1-78252-567-7. Archived (PDF) from the original on 11 February 2022. Retrieved 21 February 2022.
  103. Yee, Amy. "The country inoculating against disinformation". BBC. Archived from the original on 20 February 2022. Retrieved 21 February 2022.
  104. Sitrin, Carly (5 January 2023). "New Jersey becomes first state to mandate K-12 students learn information literacy". Politico. Archived from the original on 9 January 2023. Retrieved 9 January 2023.
  105. Roozenbeek, Jon; van der Linden, Sander; Goldberg, Beth; Rathje, Steve; Lewandowsky, Stephan (26 August 2022). "Psychological inoculation improves resilience against misinformation on social media". Science Advances. 8 (34): eabo6254. Bibcode:2022SciA....8O6254R. doi:10.1126/sciadv.abo6254. ISSN 2375-2548. PMC 9401631. PMID 36001675.
  106. "Royal Society cautions against censorship of scientific misinformation online". The Royal Society. Archived from the original on 12 February 2022. Retrieved 12 February 2022.
  107. Treen, Kathie M. d'I.; Williams, Hywel T. P.; O'Neill, Saffron J. (September 2020). "Online misinformation about climate change". WIREs Climate Change. 11 (5). Bibcode:2020WIRCC..11E.665T. doi:10.1002/wcc.665. S2CID 221879878.
  108. National Academies of Sciences, Engineering, and Medicine (22 June 2023). "Nobel Prize Summit Fuels Initiatives to Combat Misinformation and Disinformation and Build Trust in Science". National Academies of Sciences, Engineering, and Medicine. Wikidata Q124711722.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  109. Zewe, Adam. "Empowering social media users to assess content helps fight misinformation". Massachusetts Institute of Technology via techxplore.com. Archived from the original on 18 December 2022. Retrieved 18 December 2022.
  110. Jahanbakhsh, Farnaz; Zhang, Amy X.; Karger, David R. (11 November 2022). "Leveraging Structured Trusted-Peer Assessments to Combat Misinformation". Proceedings of the ACM on Human-Computer Interaction. 6 (CSCW2): 524:1–524:40. doi:10.1145/3555637. hdl:1721.1/147638.
  111. Elliott, Matt. "Fake news spotter: How to enable Microsoft Edge's NewsGuard". CNET. Archived from the original on 9 January 2023. Retrieved 9 January 2023.
  112. "12 Browser Extensions to Help You Detect and Avoid Fake News". The Trusted Web. 18 March 2021. Archived from the original on 9 January 2023. Retrieved 9 January 2023.
  113. Darcy, Oliver (24 July 2020). "Misplaced Pages administrators caution editors about using Fox News as source on 'contentious' claims | CNN Business". CNN. Archived from the original on 20 November 2021. Retrieved 9 January 2023.
  114. "New MIT Sloan research measures exposure to misinformation from political elites on Twitter". AP NEWS. 29 November 2022. Archived from the original on 18 December 2022. Retrieved 18 December 2022.
  115. Mosleh, Mohsen; Rand, David G. (21 November 2022). "Measuring exposure to misinformation from political elites on Twitter". Nature Communications. 13 (1): 7144. Bibcode:2022NatCo..13.7144M. doi:10.1038/s41467-022-34769-6. ISSN 2041-1723. PMC 9681735. PMID 36414634.
  116. ^ Nyhan B (2021). "Why the backfire effect does not explain the durability of political misperceptions". Proc Natl Acad Sci U S A. 118 (15). Bibcode:2021PNAS..11812440N. doi:10.1073/pnas.1912440117. PMC 8053951. PMID 33837144.
  117. Lazić, Aleksandra; Žeželj, Iris (18 May 2021). "A systematic review of narrative interventions: Lessons for countering anti-vaccination conspiracy theories and misinformation". Public Understanding of Science. 30 (6). SAGE Publications: 644–670. doi:10.1177/09636625211011881. ISSN 0963-6625. PMID 34006153.
  118. ^ Swire-Thompson B, DeGutis J, Lazer D (2020). "Searching for the Backfire Effect: Measurement and Design Considerations". J Appl Res Mem Cogn. 9 (3): 286–299. doi:10.1016/j.jarmac.2020.06.006. PMC 7462781. PMID 32905023.
  119. Lewandowsky, Stephan; Cook, John; Lombardi, Doug (2020), Debunking Handbook 2020, Databrary, pp. 9–11, doi:10.17910/b7.1182, retrieved 2021-01-20
  120. Guay, Brian; Berinsky, Adam J.; Pennycook, Gordon; Rand, David (August 2023). "How to think about whether misinformation interventions work". Nature Human Behaviour. 7 (8): 1231–1233. doi:10.1038/s41562-023-01667-w. ISSN 2397-3374. PMID 37563304.
  121. O'Mahony, Cian; Murphy, Gillian; Linehan, Conor (2024-11-15). "True discernment or blind scepticism? Comparing the effectiveness of four conspiracy belief interventions". Advances.in/Psychology. 2: e215691. doi:10.56296/aip00030. ISSN 2976-937X.
  122. Lawson, M. Asher; Kakkar, Hemant (2024-12-06). "Resolving conflicting findings in misinformation research: A methodological perspective". Advances.in/Psychology. 2: e235462. doi:10.56296/aip00031. ISSN 2976-937X.
  123. ^ Acerbi, Alberto (2019-02-12). "Cognitive attraction and online misinformation". Palgrave Communications. 5 (1). doi:10.1057/s41599-019-0224-y. hdl:11572/357769. ISSN 2055-1045. S2CID 257089832.
  124. ^ Almaliki, Malik (2019-04-06). "Online Misinformation Spread". Proceedings of the 2019 3rd International Conference on Information System and Data Mining. New York: ACM. pp. 171–178. doi:10.1145/3325917.3325938. ISBN 978-1-4503-6635-9. S2CID 190232270.
  125. Barthel, Michael; Mitchell, Amy; Holcomb, Jesse (2016-12-15). "Many Americans Believe Fake News Is Sowing Confusion". Pew Research Center's Journalism Project. Retrieved 2024-04-03.
  126. ^ Stawicki, Stanislaw P.; Firstenberg, Michael S.; Papadimos, Thomas J. (2020). "The Growing Role of Social Media in International Health Security: The Good, the Bad, and the Ugly". Global Health Security. Advanced Sciences and Technologies for Security Applications. pp. 341–357. doi:10.1007/978-3-030-23491-1_14. ISBN 978-3-030-23490-4. S2CID 212995901.
  127. Vosoughi, Soroush; Roy, Deb; Aral, Sinan (9 March 2018). "The spread of true and false news online" (PDF). Science. 359 (6380): 1146–1151. Bibcode:2018Sci...359.1146V. doi:10.1126/science.aap9559. PMID 29590045. S2CID 4549072. Archived from the original (PDF) on 29 April 2019. Retrieved 21 August 2019.
  128. Tucker, Joshua A.; Guess, Andrew; Barbera, Pablo; Vaccari, Cristian; Siegel, Alexandra; Sanovich, Sergey; Stukal, Denis; Nyhan, Brendan. "Social Media, Political Polarization, and Political Disinformation: A Review of the Scientific Literature". Hewlett Foundation White Paper. Archived from the original on 2019-03-06. Retrieved 2019-03-05.
  129. Machado, Caio; Kira, Beatriz; Narayanan, Vidya; Kollanyi, Bence; Howard, Philip (2019). "A Study of Misinformation in WhatsApp groups with a focus on the Brazilian Presidential Elections". Companion Proceedings of the 2019 World Wide Web Conference. pp. 1013–1019. doi:10.1145/3308560.3316738. ISBN 978-1-4503-6675-5. S2CID 153314118.
  130. ^ Allcott, Hunt; Gentzkow, Matthew; Yu, Chuan (April 2019). "Trends in the diffusion of misinformation on social media". Research & Politics. 6 (2): 205316801984855. arXiv:1809.05901. doi:10.1177/2053168019848554. S2CID 52291737.
  131. Hindman, Matthew; Davis, Trevor; Yang, Yunkang (2023-06-30). "Visual misinformation is widespread on Facebook – and often undercounted by researchers". The Conversation. Archived from the original on 2024-01-25. Retrieved 2024-01-25.
  132. ^ Chen, Xinran; Sin, Sei-Ching Joanna; Theng, Yin-Leng; Lee, Chei Sian (September 2015). "Why Students Share Misinformation on Social Media: Motivation, Gender, and Study-level Differences". The Journal of Academic Librarianship. 41 (5): 583–592. doi:10.1016/j.acalib.2015.07.003. S2CID 141523357.
  133. Shaffer, David Williamson; Collier, Wesley; Ruis, A. R. (2016). "A tutorial on epistemic network analysis: Analysing the structural connections in cognitive, social and interaction data". Journal of Learning Analytics. 3 (3): 9–45. doi:10.18608/jla.2016.33.3. ERIC EJ1126800.
  134. Milman, Oliver (2020-02-21). "Revealed: quarter of all tweets about climate crisis produced by bots". The Guardian. Archived from the original on 2020-02-22. Retrieved 2020-02-23.
  135. "Facebook's Design Makes It Unable to Control Misinformation | GW Today | The George Washington University". GW Today. Retrieved 2024-03-04.
  136. ^ Thai, My T.; Wu, Weili; Xiong, Hui (2016-12-01). Big Data in Complex and Social Networks. CRC Press. ISBN 978-1-315-39669-9.
  137. Gabbert, Fiona; Memon, Amina; Allan, Kevin; Wright, Daniel B. (September 2004). "Say it to my face: Examining the effects of socially encountered misinformation". Legal and Criminological Psychology. 9 (2): 215–227. doi:10.1348/1355325041719428. S2CID 144823646.
  138. ^ Benkler, Y. (2017). "Study: Breitbart-led rightwing media ecosystem altered broader media agenda". Archived from the original on 4 June 2018. Retrieved 8 June 2018.
  139. Törnberg, P. (2018). "Echo chambers and viral misinformation: Modeling fake news as complex contagion". PLOS ONE. 13 (9): e0203958. Bibcode:2018PLoSO..1303958T. doi:10.1371/journal.pone.0203958. PMC 6147442. PMID 30235239.
  140. Scheufele, Dietram A.; Krause, Nicole M. (16 April 2019). "Science audiences, misinformation, and fake news". Proceedings of the National Academy of Sciences. 116 (16): 7662–7669. Bibcode:2019PNAS..116.7662S. doi:10.1073/pnas.1805871115. PMC 6475373. PMID 30642953.
  141. Caramancion, Kevin Matthe (2021). "The Role of Information Organization and Knowledge Structuring in Combatting Misinformation: A Literary Analysis". Computational Data and Social Networks. Lecture Notes in Computer Science. Vol. 13116. pp. 319–329. doi:10.1007/978-3-030-91434-9_28. ISBN 978-3-030-91433-2. S2CID 244890285.
  142. ^ Starbird, Kate; Dailey, Dharma; Mohamed, Owla; Lee, Gina; Spiro, Emma S. (19 April 2018). "Engage Early, Correct More: How Journalists Participate in False Rumors Online during Crisis Events". Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. pp. 1–12. doi:10.1145/3173574.3173679. ISBN 978-1-4503-5620-6. S2CID 5046314.
  143. ^ Arif, Ahmer; Robinson, John J.; Stanek, Stephanie A.; Fichet, Elodie S.; Townsend, Paul; Worku, Zena; Starbird, Kate (2017). "A Closer Look at the Self-Correcting Crowd". Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing. pp. 155–168. doi:10.1145/2998181.2998294. ISBN 978-1-4503-4335-0. S2CID 15167363.
  144. Spradling, Matthew; Straub, Jeremy; Strong, Jay (June 2021). "Protection from 'Fake News': The Need for Descriptive Factual Labeling for Online Content". Future Internet. 13 (6): 142. doi:10.3390/fi13060142. ISSN 1999-5903.
  145. Messerole, Chris (2018-05-09). "How misinformation spreads on social media – And what to do about it". Brookings Institution. Archived from the original on 25 February 2019. Retrieved 24 February 2019.
  146. ^ Nguyen, C. Thi (2020). "Echo Chambers and Epistemic Bubbles". Episteme. 17 (2): 141–161. doi:10.1017/epi.2018.32. ISSN 1742-3600. S2CID 171520109. Archived from the original on 2022-02-09. Retrieved 2022-11-25.
  147. Bode, Leticia; Vraga, Emily K. (2 September 2018). "See Something, Say Something: Correction of Global Health Misinformation on Social Media". Health Communication. 33 (9): 1131–1140. doi:10.1080/10410236.2017.1331312. PMID 28622038. S2CID 205698884.
  148. Aral 2020, p. 43.
  149. Tucker, Emma (18 September 2022). "TikTok's search engine repeatedly delivers misinformation to its majority-young user base, report says | CNN Business". CNN. Archived from the original on 19 October 2022. Retrieved 19 October 2022.
  150. "Misinformation Monitor: September 2022". NewsGuard. Archived from the original on 18 October 2022. Retrieved 19 October 2022.
  151. Dwoskin, Elizabeth. "Misinformation on Facebook got six times more clicks than factual news during the 2020 election, study says". The Washington Post. Archived from the original on 2021-10-21. Retrieved 2021-10-22.
  152. ^ Chen, Xinran; Sin, Sei-Ching Joanna; Theng, Yin-Leng; Lee, Chei Sian (2015). "Why do Social Media Users Share Misinformation?". Proceedings of the 15th ACM/IEEE-CS Joint Conference on Digital Libraries. pp. 111–114. doi:10.1145/2756406.2756941. ISBN 978-1-4503-3594-2. S2CID 15983217.
  153. ^ Griffith, Chris (21 July 2021). "Facebook exposed over its handling of COVID misinformation". The Australian. Canberra. ProQuest 2553642687.
  154. Iyengar, Shanto; Massey, Douglas S. (16 April 2019). "Scientific communication in a post-truth society". Proceedings of the National Academy of Sciences. 116 (16): 7656–7661. Bibcode:2019PNAS..116.7656I. doi:10.1073/pnas.1805868115. PMC 6475392. PMID 30478050.
  155. ^ Swire-Thompson, Briony; Lazer, David (2 April 2020). "Public Health and Online Misinformation: Challenges and Recommendations". Annual Review of Public Health. 41 (1): 433–451. doi:10.1146/annurev-publhealth-040119-094127. PMID 31874069.
  156. "Revealed: a quarter of all tweets about climate crisis produced by bots". The Guardian. 2020-02-21. Archived from the original on 2021-04-29. Retrieved 2021-04-20.
  157. Vosoughi, Soroush; Roy, Deb; Aral, Sinan (2018-03-09). "The spread of true and false news online". Science. 359 (6380): 1146–1151. Bibcode:2018Sci...359.1146V. doi:10.1126/science.aap9559. ISSN 0036-8075. PMID 29590045. S2CID 4549072.
  158. Shin, Jieun; Jian, Lian; Driscoll, Kevin; Bar, François (June 2018). "The diffusion of misinformation on social media: Temporal pattern, message, and source". Computers in Human Behavior. 83: 278–287. doi:10.1016/j.chb.2018.02.008. S2CID 41956979.
  159. "Amazon to suspend Parler after deadly Capitol Hill riot". Al Jazeera. 10 January 2021. Archived from the original on 7 March 2022. Retrieved 7 March 2022.
  160. Landrum, Asheley R.; Olshansky, Alex (2019). "The role of conspiracy mentality in denial of science and susceptibility to viral deception about science". Politics and the Life Sciences. 38 (2): 193–209. doi:10.1017/pls.2019.9. ISSN 0730-9384. PMID 32412208.
  161. ^ Diaz Ruiz, Carlos; Nilsson, Tomas (2023). "Disinformation and Echo Chambers: How Disinformation Circulates on Social Media Through Identity-Driven Controversies". Journal of Public Policy & Marketing. 42 (1): 18–35. doi:10.1177/07439156221103852. ISSN 0743-9156.
  162. Allgaier, Joachim (July 25, 2019). "Science and Environmental Communication on YouTube: Strategically Distorted Communications in Online Videos on Climate Change and Climate Engineering". Front. Commun. 4 (4): 36. doi:10.3389/fcomm.2019.00036.
  163. ^ Nugent, Ciara (January 16, 2020). "YouTube Has Been 'Actively Promoting' Videos Spreading Climate Denialism, According to New Report". Time. Archived from the original on January 31, 2020.
  164. ^ Paul, Kari (24 November 2020). "OANN suspended from YouTube after promoting a sham cure for Covid-19". The Guardian. Archived from the original on 25 November 2020. YouTube's Covid-specific misinformation policies prohibit content that disputes the existence of the virus, discourages someone from seeking medical treatment for Covid, disputes guidance from local health authorities on the pandemic, or offers unsubstantiated medical advice or treatment.
  165. "Sky News Australia barred for week by YouTube over Covid misinformation". BBC News. 2021-08-01. Retrieved 2021-08-01.
  166. De Vynck, Gerrit (September 28, 2021). "YouTube is banning prominent anti-vaccine activists and blocking all anti-vaccine content". The Washington Post. Archived from the original on September 29, 2021.
  167. Wakabayashi, Daisuke; Hsu, Tiffany (October 7, 2021). "Google bans ads on content, including YouTube videos, with false claims about climate change". The New York Times. Archived from the original on October 8, 2021.
  168. Mishra, Stuti (May 4, 2023). "Google profiting from climate misinformation on YouTube, report finds". The Independent. Archived from the original on 27 August 2023.
  169. ^ "YouTube making money off new breed of climate denial, monitoring group says". Reuters. 16 January 2024. Archived from the original on 16 January 2024.
  170. Stapleton, Paul (2003). "Assessing the quality and bias of web-based sources: implications for academic writing". Journal of English for Academic Purposes. 2 (3): 229–245. doi:10.1016/S1475-1585(03)00026-2.
  171. West, Jevin D.; Bergstrom, Carl T. (2021). "Misinformation in and about science". Proceedings of the National Academy of Sciences. 118 (15). Bibcode:2021PNAS..11812444W. doi:10.1073/pnas.1912444117. PMC 8054004. PMID 33837146.
  172. ^ "Facebook's Lab-Leak About-Face". The Wall Street Journal. Archived from the original on 2021-07-14. Retrieved 2021-07-14.
  173. "Covid origin: Why the Wuhan lab-leak theory is being taken seriously". BBC News. 27 May 2021. Archived from the original on 30 June 2021. Retrieved 14 July 2021.
  174. "Hydroxychloroquine: Why a video promoted by Trump was pulled on social media". BBC News. 2020-07-28. Archived from the original on 2020-11-02. Retrieved 2021-11-24.
  175. "Stella Immanuel – the doctor behind unproven coronavirus cure claim". BBC News. 2020-07-29. Archived from the original on 2021-10-11. Retrieved 2020-11-23.
  176. Bertrand, Natasha (October 19, 2020). "Hunter Biden story is Russian disinfo, dozens of former intel officials say". Politico. Archived from the original on October 20, 2020. Retrieved October 20, 2020.
  177. Lizza, Ryan (September 21, 2021). "Politico Playbook: Double Trouble for Biden". Politico. Archived from the original on November 23, 2021. Retrieved September 23, 2021.
  178. Gladstone, Brooke (2012). The Influencing Machine. New York: W. W. Norton & Company. pp. 49–51. ISBN 978-0-393-34246-8.
  179. Stewart, Mallory (2021). "Defending Weapons Inspections from the Effects of Disinformation". AJIL Unbound. 115: 106–110. doi:10.1017/aju.2021.4. S2CID 232070073.
  180. Brosnan, Deanne (13 January 2021). "When Misinformation is Misinformation". CE Think Tank Newswire. Miami. ProQuest 2477885938.
  181. ^ "Deepfakes, Elections, and Shrinking the Liar's Dividend | Brennan Center for Justice". www.brennancenter.org. 2024-02-08. Retrieved 2024-04-12.
  182. Bodner, Glen E.; Musch, Elisabeth; Azad, Tanjeem (December 2009). "Reevaluating the potency of the memory conformity effect". Memory & Cognition. 37 (8): 1069–1076. doi:10.3758/MC.37.8.1069. PMID 19933452.
  183. Southwell, Brian G.; Thorson, Emily A.; Sheble, Laura (2018). Misinformation and Mass Audiences. University of Texas Press. ISBN 978-1-4773-1458-6.
  184. Barker, David (2002). Rushed to Judgement: Talk Radio, Persuasion, and American Political Behavior. New York: Columbia University Press. pp. 106–109.
  185. "The misinformation that was told about Brexit during and after the referendum". The Independent. 2 August 2018. Archived from the original on 15 May 2022.
  186. Marwick, Alice; Lewis, Rebecca (2017). Media Manipulation and Disinformation Online. New York: Data & Society Research Institute. pp. 40–45.
  187. Aral 2020, pp. 45–47.
  188. O'Connor, Cailin; Weatherall, James Owen (2019). The Misinformation Age: How False Beliefs Spread. New Haven: Yale University Press. pp. 10. ISBN 978-0-300-23401-5.
  189. Sinha, P.; Shaikh, S.; Sidharth, A. (2019). India Misinformed: The True Story. Harper Collins. ISBN 978-93-5302-838-1.
  190. ^ Bratu, Sofia (May 24, 2020). "The Fake News Sociology of COVID-19 Pandemic Fear: Dangerously Inaccurate Beliefs, Emotional Contagion, and Conspiracy Ideation". Linguistic and Philosophical Investigations. 19: 128–134. doi:10.22381/LPI19202010.
  191. Gayathri Vaidyanathan (22 July 2020). "News Feature: Finding a vaccine for misinformation". Proceedings of the National Academy of Sciences of the United States of America. 117 (32): 18902–18905. Bibcode:2020PNAS..11718902V. doi:10.1073/PNAS.2013249117. ISSN 0027-8424. PMC 7431032. PMID 32699146. Wikidata Q97652640.
  192. "Misinformation on coronavirus is proving highly contagious". AP NEWS. 2020-07-29. Archived from the original on 2020-11-20. Retrieved 2020-11-23.
  193. "Info-Environmentalism: An Introduction". Archived from the original on 2018-07-03. Retrieved 2018-09-28.
  194. "Information Environmentalism". Digital Learning and Inquiry (DLINQ). 2017-12-21. Archived from the original on 2018-09-28. Retrieved 2018-09-28.
  195. Righetti, Nicola; Rossi, Luca; Marino, Giada (4 July 2022). "At the onset of an infodemic: Geographic and disciplinary boundaries in researching problematic COVID-19 information". First Monday. doi:10.5210/fm.v27i7.12557. S2CID 250289817.
  196. Bernstein, Joseph (9 August 2021). "Bad News: Selling the story of disinformation". Harper's Magazine. Archived from the original on 26 September 2022. Retrieved 30 September 2022.
  197. Adler-Bell, Sam (2022-05-20). "The Liberal Obsession With 'Disinformation' Is Not Helping". Intelligencer. Archived from the original on 2022-09-30. Retrieved 2022-09-30.
  198. Camargo, Chico Q.; Simon, Felix M. (20 September 2022). "Mis- and disinformation studies are too big to fail: Six suggestions for the field's future". Harvard Kennedy School Misinformation Review. doi:10.37016/mr-2020-106. S2CID 252423678.
  199. Westerlund, Mika (2019-01-01). "The Emergence of Deepfake Technology: A Review". Technology Innovation Management Review. 9 (11): 39–52. doi:10.22215/timreview/1282. ISSN 1927-0321.
  200. Ferrara, Emilio (2017-07-31). "Disinformation and Social Bot Operations in the Run Up to the 2017 French Presidential Election". First Monday. arXiv:1707.00086. doi:10.5210/fm.v22i8.8005. ISSN 1396-0466.
  201. Bakshy, Eytan; Messing, Solomon; Adamic, Lada A. (2015-06-05). "Exposure to ideologically diverse news and opinion on Facebook". Science. 348 (6239): 1130–1132. Bibcode:2015Sci...348.1130B. doi:10.1126/science.aaa1160. ISSN 0036-8075. PMID 25953820.
  202. "Document-level Claim Extraction and Decontextualisation for Fact-Checking". arxiv.org. Retrieved 2024-10-10.
  203. Afchar, Darius; Nozick, Vincent; Yamagishi, Junichi; Echizen, Isao (December 2018). "MesoNet: A Compact Facial Video Forgery Detection Network". 2018 IEEE International Workshop on Information Forensics and Security (WIFS). IEEE. pp. 1–7. arXiv:1809.00888. doi:10.1109/wifs.2018.8630761. ISBN 978-1-5386-6536-7.
  204. Carroll, Alexander J.; Borycz, Joshua (July 2024). "Integrating large language models and generative artificial intelligence tools into information literacy instruction". The Journal of Academic Librarianship. 50 (4): 102899. doi:10.1016/j.acalib.2024.102899.

Further reading

Library resources about
Misinformation

External links

Disinformation and misinformation
Types
Books and documentaries
Examples
Health
Operations and events by country
CanadaJihadunspun.com
China
CzechoslovakiaOperation Neptune
Germany
India
IsraelTeam Jorge
KoreaVoluntary Agency Network of Korea
KuwaitFintas Group
MexicoPeñabots
Philippines
Russia /
Soviet Union
Soviet era
Post-Soviet era
South AfricaHIV/AIDS denialism
Turkey
United Kingdom
United States
VenezuelaBolivarian Army of Trolls
Vietnam
Opposition
Categories: