David,
I see your point, but the problem with treating “disinformation” as undefinable is that it concedes the absence of objective reality. Yes, both sides use it to accuse the other of lying, but in many cases it’s possible to determine that one side actually is making false statements, and that the false statements are made with the intent of convincing others to believe “facts” that are objectively false. (And, in many cases, to then use those false beliefs to justify actions in furtherance of an agenda that is untethered to the truth of the underlying statements.) To reduce that to “political name-calling” is a form of false equivalence.
Michelle
(speaking only for myself)
From: osi20...@googlegroups.com <osi20...@googlegroups.com>
On Behalf Of David Wojick
Sent: Tuesday, December 14, 2021 3:59 PM
To: Biagioli, Mario <biag...@law.ucla.edu>
Cc: Glenn Hampson <gham...@nationalscience.org>; The Open Scholarship Initiative <osi20...@googlegroups.com>
Subject: Re: Don't share links on social media
It means there is no way to tell what it applies to. In analytic philosophy a paradigm case of conceptual confusion is when there is disagreement at to what a term applies to. In this case that lack of agreement is virtually complete. The term is universally used as a form of political name calling, where the other side disagrees completely.
David
On Dec 14, 2021, at 2:49 PM, Biagioli, Mario <biag...@law.ucla.edu> wrote:
David,
What do you mean by misinformation being a "perfectly incoherent" concept?
Mario
From:
osi20...@googlegroups.com <osi20...@googlegroups.com> on behalf of David Wojick <dwo...@craigellachie.us>
Sent: Tuesday, December 14, 2021 10:22 AM
To: Glenn Hampson <gham...@nationalscience.org>
Cc: The Open Scholarship Initiative <osi20...@googlegroups.com>
Subject: Re: Don't share links on social media
"Don’t share links to news articles on social media." is worse than ridiculous. Most of what I do on social media is discuss news articles, including this one! Should this email, and yours, have been prohibited?
As for"misinformation" this concept is typically used in political debates, especially scientifically intensive policy debates, to attack one's opponents. It is as perfectly incoherent as concepts get.
I will defend to the death people's right to say flakey science. (Apologies to Voltaire.)
David
On Dec 14, 2021, at 12:58 PM, Glenn Hampson <gham...@nationalscience.org> wrote:
Jason’s History Club discussed a provocative idea a few days ago regarding how to curb history misinformation: Don’t share links to news articles on social media.
Can we extend this same thinking to research as well? At first blush, sharing less information seems utterly opposed to the goals of openness. But is it really? Making information discoverable and accessible isn’t the same as promoting flaky science on Twitter. Is there a best practices line here between openness and promotion that we should define, in the interest of nurturing openness but at the same time not nurturing misinformation?
From: Jason Steinhauer from History Club <jasonst...@substack.com>
Sent: Sunday, December 12, 2021 5:01 AM
To: gcha...@gmail.com
Subject: Don't share links on social media
In the fight against disinformation online, we have a powerful tool in our toolbox
Don't share links on social mediaIn the fight against disinformation online, we have a powerful tool in our toolbox
This week we held six online conversations about my new book History, Disrupted: five on Clubhouse and one on Twitter Spaces. (Want to listen to the podcast recordings? Click here, here, here, here, here & here.) In each conversation I was asked for one concrete action we can take to stem the flow of misinformation and disinformation online. My response: don’t share links to news articles on social media. This radical suggestion was met by agreement among some participants and skepticism and pushback from others. So, let me briefly pull on this thread in this week’s newsletter, laying the groundwork for future conversations on the subject. A major theme throughout History, Disrupted is how and why social media incentivizes certain behaviors that lead to particular outcomes. Take virality, as one example. Because the media and social media ecosystems reward virality with influence—money, fame, power, and attention—they incentivize actors with an array of agendas to purposefully design content that will go viral. In my book, I devote an entire chapter to what I call, “the viral past,” a type of online content that uses information about history to send information virally through networks as a means to advance agendas. Those who can make the past go viral get rewarded by the news media and social media with power and influence; those who cannot never have their ideas reach a wider public, even if those ideas contribute meaningfully to public understandings of history. In conversations this week, I argued that this is a flawed incentive structure. When people are rewarded for attention and not accuracy, the tail of gaining influence online can wind up wagging the dog of trying to discern with thoughtfulness and accuracy what may actually be occurring in the world—or, in the case of history, what may have actually occurred in the past. When we incentivize clicks, views and shares, the accuracy of information can become a secondary consideration. When virality becomes a pathway to power, actors who are fighting for power have shown that they will chase it at all costs. That can include political parties, corporate media, activists, and disinformation agents. It is us, however, who collectively make this possible. Through the mass sharing of links to political news stories on our social media pages, we create the conditions for virality and misinformation to thrive. In essence, we do the work of the political parties, media companies, and disinformation agents for them, using our networks and weak ties to spread ideas from one cluster of social media users to another. In the case of political parties and corporate media who have millions, if not billions, of dollars at their disposal, we do this work for them without receiving any compensation. What we get is an emotion: a sense of virtue, a sense of outrage, an opportunity to champion our values or beliefs, or a feeling that we’ve contributed to pressing political debates. Sharing becomes an act of self-communication, communicating to our followers and friends where we stand on a particular issue. The past two decades have shown just how animating that desire to personify our virtues can be. So, how do we begin to nip this in the bud? We stop sharing. Without our networks and connections, political news content cannot flow from one cluster of users to another. Without our networks and connections, the engine of virality loses its steam. Here is how I articulated it to Maceo Paisley in our conversation on Clubhouse this past Friday:
This approach, of course, raises numerous questions. How can we support journalists and journalism if not through sharing their articles on social media? (Ideas include subscribing to their publications, following particular reporters you appreciate, and hosting social audio discussions about items in the news on platforms such as Clubhouse and Twitter Spaces.) Can we share links to other things on our social media pages? (Yes! Recipes, cat videos, and GoFundMe pages are examples of great things to share.) Will this stop all misinformation and disinformation? (No, but it might make a dent.) What if I really want to share news and political views with my family and friends? (There are other ways). To the last point, email continues to be a very effective way to share news articles with friends and colleagues. My mother has been emailing me news articles since 2004. She does not have social media (thank goodness), but that has not prevented her from expressing her political views or participating actively in our democracy. Other forms of discourse are possible. This is, obviously, a larger conversation that cannot be given justice in a single newsletter. But the conversation feels particularly urgent as the contours of Web 3.0 begin to take shape. If we look back on the past twenty years of Web 2.0, we can likely agree that it did not turn out to be the panacea it was initially promised. In part, that was because we were not as intentional or thoughtful as we needed to be about how social media would work, what incentive structures were being created, and what the effects might be on history, journalism, politics and democracy. Similar halcyon visions of Web 3.0 are currently being marketed to us, a future of the Web that feels both exciting and perilous. We must be intentional about designing the right incentives into these new platforms, lest we look back in twenty years and wish that we had. A radical re-thinking and re-examining may be required. Squashing the urge to share might be one of them. Thank you’s & shout-outsThere are many people to thank for the amazing conversations and support of History, Disrupted this week, among them:
Book updateFor those who ordered History, Disrupted: How Social Media & the World Wide Web Have Changed the Past on Amazon, you may have received an email this week saying that your shipment has been delayed. Apparently the book is temporarily out of stock; I guess we sold more copies than they were prepared for! The publisher is aware and printing more copies. You should receive an update soon on when the book will ship. The book is also available through Barnes & Noble and at local independent book sellers, including Lost City Books in Washington, D.C. For those in the D.C. area, we’ll be having a book launch celebration at Lost City Books on January 7, 2022 at 7:30 p.m. More information to come. Have a good week. History Club meets Thursdays at 10 pm ET. Want to participate? Suggest a topic for a future conversation. Haven’t supported the History Club? Please consider it. Your support allows me to publish posts like this. You’re a free subscriber to The History Club. For the full experience, become a paid subscriber.
© 2021 Jason Steinhauer
Unsubscribe |
--
As a public and publicly-funded effort, the conversations on this list can be viewed by the public and are archived. To read this group's complete listserv policy (including disclaimer and reuse information), please visit
http://osinitiative.org/osi-listservs.
---
You received this message because you are subscribed to the Google Groups "The Open Scholarship Initiative" group.
To unsubscribe from this group and stop receiving emails from it, send an email to
osi2016-25+...@googlegroups.com.
To view this discussion on the web visit
https://groups.google.com/d/msgid/osi2016-25/013d01d7f10b%24e0097750%24a01c65f0%24%40nationalscience.org.
--
As a public and publicly-funded effort, the conversations on this list can be viewed by the public and are archived. To read this group's complete listserv policy (including disclaimer and reuse information), please visit
http://osinitiative.org/osi-listservs.
---
You received this message because you are subscribed to the Google Groups "The Open Scholarship Initiative" group.
To unsubscribe from this group and stop receiving emails from it, send an email to
osi2016-25+...@googlegroups.com.
To view this discussion on the web visit
https://groups.google.com/d/msgid/osi2016-25/0762475A-7405-45F1-A9EE-7581BB99953C%40craigellachie.us.
--
As a public and publicly-funded effort, the conversations on this list can be viewed by the public and are archived. To read this group's complete listserv policy (including disclaimer and reuse information), please visit
http://osinitiative.org/osi-listservs.
---
You received this message because you are subscribed to the Google Groups "The Open Scholarship Initiative" group.
To unsubscribe from this group and stop receiving emails from it, send an email to
osi2016-25+...@googlegroups.com.
To view this discussion on the web visit
https://groups.google.com/d/msgid/osi2016-25/BYAPR13MB42805FADD9BC31FBAE87E7A9E1759%40BYAPR13MB4280.namprd13.prod.outlook.com.
--
As a public and publicly-funded effort, the conversations on this list can be viewed by the public and are archived. To read this group's complete listserv policy (including disclaimer and reuse information), please visit
http://osinitiative.org/osi-listservs.
---
You received this message because you are subscribed to the Google Groups "The Open Scholarship Initiative" group.
To unsubscribe from this group and stop receiving emails from it, send an email to
osi2016-25+...@googlegroups.com.
To view this discussion on the web visit
https://groups.google.com/d/msgid/osi2016-25/697CFBDF-7A78-4B55-BBF5-48844E943B33%40craigellachie.us.
The terms are largely synonymous. False statements intended to deceive/mislead. If there is a difference, it is one of nuance – disinformation is false statements known to be false and intended to deceive, “misinformation” is often used to include the possibility that the speaker is mistaken rather than lying, while still intending the statement to be believed and influence behavior.
I respectfully disagree that the term(s) is/are meaningless in the political context just because both sides use it. Facts do exist, and false factual statements made intentionally for political purposes are still false.
I agree with you, however, that using social media but not linking to news stories doesn’t seem to be very effective. Better to just not use social media.
Regards,
Michelle
I was recently a subject in a study being conducted this fall by a research group in the ISchool at the University of Washington, the focus of which was to:
study the way information professionals talk about or experience problems related to false and misleading information (misinformation, disinformation, etc.) in the field.
After I had answered the questions posed by the researcher, he then gave me a little more detail about their work, and one of the items that came up (in part because I differentiated between the two in one of my answers) was misinformation vs. disinformation. He indicated that researchers in the field DO differentiate those two terms, and that the differentiation is an important one. What the researcher basically said (sorry I don’t have it verbatim because this was all verbal in a Zoom session) was that:
Misinformation is false / misleading /out-of-context information that is spread by ignorance or because the person spreading it believes it to be true. With misinformation there is NO INTENT to deceive.
Disinformation is false / misleading / out-of-context information this is spread when the individual KNOWS the information is untrue. With disinformation there is INTENT to deceive.
Mel
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Mel DeSart (he/him/his)
Head, Engineering Library and Head, Mathematics Research Library
University of Washington
Box 352170 des...@uw.edu
Seattle, WA 98195-2170 office: 206-685-8369
“It is not written in the stars that I will always understand what is going
on -- a truism that I often find damnably annoying."
Robert A. Heinlein, from his novel "Friday"
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
From: osi20...@googlegroups.com <osi20...@googlegroups.com>
On Behalf Of David Wojick
Sent: Tuesday, December 14, 2021 3:41 PM
To: Gluck, Michelle <mvg...@psu.edu>
To view this discussion on the web visit https://groups.google.com/d/msgid/osi2016-25/F53CEA3B-596A-4D7A-B76D-C5F6753F79FB%40craigellachie.us.
Anthony
To unsubscribe from this group and stop receiving emails from it, send an email to unsub...@googlegroups.com">osi2016-25+unsub...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/osi2016-25/013d01d7f10b%24e0097750%24a01c65f0%24%40nationalscience.org.
--
As a public and publicly-funded effort, the conversations on this list can be viewed by the public and are archived. To read this group's complete listserv policy (including disclaimer and reuse information), please visit http://osinitiative.org/osi-listservs.
---
You received this message because you are subscribed to the Google Groups "The Open Scholarship Initiative" group.
To unsubscribe from this group and stop receiving emails from it, send an email to unsub...@googlegroups.com">osi2016-25+unsub...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/osi2016-25/0762475A-7405-45F1-A9EE-7581BB99953C%40craigellachie.us.
--
As a public and publicly-funded effort, the conversations on this list can be viewed by the public and are archived. To read this group's complete listserv policy (including disclaimer and reuse information), please visit http://osinitiative.org/osi-listservs.
---
You received this message because you are subscribed to the Google Groups "The Open Scholarship Initiative" group.
To unsubscribe from this group and stop receiving emails from it, send an email to unsub...@googlegroups.com">osi2016-25+unsub...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/osi2016-25/BYAPR13MB42805FADD9BC31FBAE87E7A9E1759%40BYAPR13MB4280.namprd13.prod.outlook.com.
--
As a public and publicly-funded effort, the conversations on this list can be viewed by the public and are archived. To read this group's complete listserv policy (including disclaimer and reuse information), please visit http://osinitiative.org/osi-listservs.
---
You received this message because you are subscribed to the Google Groups "The Open Scholarship Initiative" group.
To unsubscribe from this group and stop receiving emails from it, send an email to unsub...@googlegroups.com">osi2016-25+unsub...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/osi2016-25/697CFBDF-7A78-4B55-BBF5-48844E943B33%40craigellachie.us.
--
As a public and publicly-funded effort, the conversations on this list can be viewed by the public and are archived. To read this group's complete listserv policy (including disclaimer and reuse information), please visit http://osinitiative.org/osi-listservs.
---
You received this message because you are subscribed to the Google Groups "The Open Scholarship Initiative" group.
To unsubscribe from this group and stop receiving emails from it, send an email to unsub...@googlegroups.com">osi2016-25+unsub...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/osi2016-25/PH0PR02MB7303DE6C16733AC96A1589C0AA759%40PH0PR02MB7303.namprd02.prod.outlook.com.
On Dec 14, 2021, at 6:11 PM, Mel DeSart <des...@uw.edu> wrote:
To view this discussion on the web visit https://groups.google.com/d/msgid/osi2016-25/BY3PR08MB721996F8E644B93F1E7E9A45B7759%40BY3PR08MB7219.namprd08.prod.outlook.com.
No David, you said
that is someone asserting what they do not in fact believe
Belief is largely irrelevant. If what you meant by the phrase above is someone asserting as true what they know to be false, then I stand corrected. But that’s not really what the phrase above says.
A politician who does not believe in the efficacy of the COVID-19 vaccine can still (in an effort to secure votes) tell voters that the vaccines work and that people should get vaccinated, even though they don’t actually believe that. That is of course lying on the part of the politician, but it’s not actually spreading disinformation because what the politician is saying is true, not false, even though the politician doesn’t _believe_ it to be true.
Mel