• Nie Znaleziono Wyników

Three contextual dimensions of information on social media

N/A
N/A
Protected

Academic year: 2021

Share "Three contextual dimensions of information on social media"

Copied!
11
0
0

Pełen tekst

(1)

Delft University of Technology

Three contextual dimensions of information on social media

lessons learned from the COVID-19 infodemic

Marin, L. DOI 10.1007/s10676-020-09550-2 Publication date 2020 Document Version Final published version Published in

Ethics and Information Technology

Citation (APA)

Marin, L. (2020). Three contextual dimensions of information on social media: lessons learned from the COVID-19 infodemic. Ethics and Information Technology. https://doi.org/10.1007/s10676-020-09550-2

Important note

To cite this publication, please use the final published version (if applicable). Please check the document version above.

Copyright

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons. Takedown policy

Please contact us and provide details if you believe this document breaches copyrights. We will remove access to the work immediately and investigate your claim.

This work is downloaded from Delft University of Technology.

(2)

1 23

Ethics and Information Technology

ISSN 1388-1957 Ethics Inf Technol

DOI 10.1007/s10676-020-09550-2

Three contextual dimensions of

information on social media: lessons

learned from the COVID-19 infodemic

(3)

1 23

Your article is published under the Creative

Commons Attribution license which allows

users to read, copy, distribute and make

derivative works, as long as the author of

the original work is cited. You may

self-archive this article on your own website, an

institutional repository or funder’s repository

and make it publicly available immediately.

(4)

Vol.:(0123456789)

1 3

Ethics and Information Technology

https://doi.org/10.1007/s10676-020-09550-2

ORIGINAL PAPER

Three contextual dimensions of information on social media: lessons

learned from the COVID‑19 infodemic

Lavinia Marin1

© The Author(s) 2020

Abstract

The COVID-19 pandemic has been accompanied on social media by an explosion of information disorders such as inaccurate, misleading and irrelevant information. Countermeasures adopted thus far to curb these informational disorders have had limited success because these did not account for the diversity of informational contexts on social media, focusing instead almost exclusively on curating the factual content of user’s posts. However, content-focused measures do not address the primary causes of the infodemic itself, namely the user’s need to post content as a way of making sense of the situation and for gathering reactions of consensus from friends. This paper describes three types of informational context—weak epistemic, strong normative and strong emotional—which have not yet been taken into account by current measures to curb down the informational disorders. I show how these contexts are related to the infodemic and I propose measures for dealing with them for future global crisis situations.

Keywords Disinformation · Infodemic · Covid-19 · Designed interaction · Online emotions · Context sensitive design

Informational disorders on social media

relating to COVID‑19

From the beginning, the pandemic was reflected in the online realm of social media by a flood of redundant infor-mation (the so-called infodemic) out of which a significant percentage was made up of misinformation and disinfor-mation (MDI). MDI is used here as one umbrella term for two distinct phenomena: misinformation, which is usually defined as false information shared without knowledge that it is false, while disinformation is fabricated information distributed with the clear intention to mislead (Fallis 2014, p.136). However, on social media, these distinctions are hard to maintain sharply, since we cannot always know who cre-ated a piece of misleading information and with what pur-pose. The infodemic is understood as “an overabundance of information—some accurate and some not—that makes it hard for people to find trustworthy sources and reliable guid-ance when they need it” (World Health Organization 2020). This concept is similar with that of information pollution

which was defined as “irrelevant, redundant, unsolicited and low-value information” (Wardle and Derakhshan 2018, p.10) but applied to a crisis situation when an infodemic can become dangerous (Tangcharoensathien et al. 2020, p. 2).

Already since February 2020, the World Health Organi-zation had singled the infodemic as an emerging problem in the context of the COVID-19 pandemic. One of the most problematic aspects of an infodemic is that it creates infor-mation overload which leads to inforinfor-mation fatigue for online users: the user’s capacity for paying attention to infor-mation is limited and tends to exhaust quickly. Studies in the psychology of social media have shown that, under condi-tions of informational overload, users will revert to using mental shortcuts or heuristics for assessing new information. By employing cognitive heuristics, social media users tend to rely on their friends and their endorsements in selecting what information to trust or engage with (Koroleva et al. 2010, p. 5). Taking these mental shortcuts occurs “in condi-tions of low motivation and limited ability to process the incoming information” (Koroleva et al. 2010, p. 5) which is arguably the case when dealing with information overload. Trying to save their mental energy, users will tend to del-egate their critical engagement to their social network, trust-ing their peers, or to become disengaged from so many news and revert to apathy. Both strategies are dangerous because * Lavinia Marin

l.marin@tudelft.nl

1 Ethics and Philosophy of Technology Section, TU Delft,

(5)

L. Marin

1 3

social media users are also citizens who are instrumental in the efforts to curb down the pandemic. If citizens do not correctly understand what they have to do in a pandemic situation, then governmental measures will be ineffective.

Online information about the COVID-19 pandemic posted on social media displayed two seemingly distinct problems: the rapid propagation of misinformation and dis-information (MDI) and the so-called infodemic. Both prob-lems concerned how information travels in an online social networking medium, but only one of them was tackled with some degree of efficiency. Social media platforms rapidly stepped up their pre-existing measures of dealing with MDI and targeted specifically the COVID-19 related misleading information, whereas the infodemic remained untouched (Howard 2020). The infodemic was seen as a side-effect of the intensification of user interactions on social media, some informational noise that accompanied the humming of online communications.

In this article, I will argue that both informational lems are related with and symptomatic of a deeper prob-lem embedded in social media: the contextual design of the user’s interactions with information. While we certainly need to pay attention to the quality of informational con-tent (Floridi and Illari 2014) distributed on social media, similar attention should be paid to the quality of the user’s interaction with the informational context. Measures focus-ing solely on the factual content of information distributed online risk ignoring a significant aspect of social media, its particular context of engaging with information. The paper is structured as follows: first, I briefly review the measures taken by social media platforms to deal with the COVID-19 related MDI, classifying them in view of the content or context focus. Secondly, I describe three dimensions of the informational context on social media which made MDI par-ticularly difficult to deal with during the pandemic and, by extension, aggravated the infodemic. Finally, drawing from the contextual approach to information, I propose some measures to tackle informational disorders online for future similar global crisis situations.

Counter‑measures against mis‑/

dis‑information related to COVID‑19

on social media

MDI related to COVID-19 was tackled visibly by most main-stream social media platforms. This was possible because there were already some methods in place for dealing with MDI. Starting with the 2016 elections and the Cambridge Analytica scandal, social media platforms began paying attention to the MDI shared by their users. In recent years, social media platforms have been testing methods of con-tent curating by using external fact-checking organisations,

flagging the misinforming content, and sometimes removing it (Howard 2020). In the wake of the COVID-19 pandemic, these efforts for fact-checking were accelerated to an impres-sive extent: “the number of English-language fact-checks rose more than 900% from January to March” (Brennen et al. 2020). This concentration of effort from disparate organisa-tions was motivated by the emergency of the situation but also by the topic of COVID-19 which rendered itself easier to fact-check: it was more or less clear what was false con-tent1 and what not—as opposed to previous cases of political

MDI (Brennen et al. 2020).

Table 1 below summarises the main measures taken by social media platforms for dealing with MDI during the pandemic.2

Most approaches listed in the previous table were pri-marily content-focused as they were targeting the factual or descriptive content of the information, by checking it against existing evidence and reducing its visibility. However, the context of MDI is just as important as its content (Wardle and Derakhshan 2018). Contextual MDI may appear when placing genuine information in a fabricated setting, for example, a private statement cited as if it were a public one, a personal opinion as to represent the views of an organisa-tion, mixing facts with irrelevant comments, or changing the date or place where a photo was taken. A large extent of COVID-19 related MDI shared on social media was con-textual or, as some have put it, reconfigured: “most (59%) of the misinformation in our sample involves various forms of reconfiguration, where existing and often true informa-tion is spun, twisted, recontextualised, or reworked. Less misinformation (38%) was completely fabricated” (Brennen et al. 2020). Even those measures which combined content with context awareness (such as numbers 3,5,6 and 7 in the Table 1), the content was the primary focus. These measures relied on someone checking first what information was good enough and then proposing some contextual measures.

Content-focused approaches are not misguided, but tend to give most of the agency to the social media platform, while the users are left with a passive role, to click and react to what they are shown. Meanwhile, when a meas-ure is both content and context focused, the user’s agency starts to play a significant role. No matter how many alter-native sources of information one is exposed to (as it is the case with measures 5,6,7 in the Table 1), the user has still the final choice to engage with them or not. By contrast, a

1 For a list of categories of false claims about Covid-19, see the

dynamic graph found at https ://euvsd isinf o.eu/throw ing-coron aviru s-disin fo-at-the-wall-to-see-what-stick s/.

2 Data sources used for this table: Silverman (April 2019), Naeem

and Bhatti (2020), Cinelli et  al. (2020), Tangcharoensathien et  al. (2020), EU vs Disinfo. (2020) and Brennen et al. (2020).

(6)

Three contextual dimensions of information on social media: lessons learned from the COVID-19…

1 3

context-focused approach would presuppose that the user has the freedom to use the social media platform in a way that stimulates and recognises other kinds of engagement with information, regardless whether the information at hand is genuine or MDI. A context-focused approach cre-ates informational environments which accommodate the possibility that the users may engage differently with and interpret differently the same information. This approach would aim to stimulate the users in becoming more sensi-tive to the modes in which the information is presented to them, educating them in the long run.

This notion of context-sensitivity was inspired by similar constructs encountered in Human–Computer Interaction studies such as “contextual design” (Wixon et al. 1990) and “context-aware computing” (Schilit et al. 1994–1994). Context sensitivity in design starts from the idea that the user’s perspective is not that predictable, and that there is not one single context of use, accepting that users make sense of an interface in various ways, hence proposing that the designer accommodates for multiple meanings and complex interactions right from the start. In the long run, a contextual approach can educate the users without trying to change their beliefs or attitudes about the informational content as such. The contextual information approach is inspired by Value Sensitive Design (Friedman 1996; van den Hoven 2007) which is complemented by paying attention to three particular dimensions of infor-mational context on social media, described in the next section.

Three dimensions of the contextual

information on social media

The strong emotional context

Before the pandemic, it was already shown that MDI propa-gates on social media platforms by playing on the emotional reactions of the online audience (Zollo et al. 2015; Khal-darova and Pantti 2016), aiming to deliberately stir power-ful emotions in their readers. Some researchers called this feature of MDI a form of “empathic optimisation” (Bakir and McStay 2018, p. 155). Emotional manipulation in news items (especially click-bait) is an efficient way of capturing user’s attention since emotion-stirring news are usually more interacted with than the neutral ones (Bakir and McStay 2018, p. 155). It may seem that only MDI is emotionally loaded, whereas genuine news sound more sober and neutral. But this would be a misleading view of how information travels on social media platforms. Emotional reactions do not belong to misleading information alone, rather these are a normal side-effect of the emotional infrastructure already embedded in most social media platforms.

Users of social media platforms are allowed a palette of actions and reactions: some are seemingly neutral (com-menting, sharing and posting) while others have a clear emo-tional valence: liking and using other emoticons to endorse or dislike a post. These emotionally charged reactions are easier to perform than the neutral ones: it takes a split second to click like on a post, but some more time to comment on

Table 1 Main types of measures against MDI on social media during the pandemic

No. Counter-measures Approach to MDI Type of approach Platforms implementing it

1 Flagging misinformation by regular users—double-checked by editors later

Decreases the quantity of MDI Content-focused Facebook, Twitter 2 Fact-checking by certified third

parties Decreases the quantity of MDI, hence the user’s interaction with it

Content-focused Facebook, Twitter 3 Less visibility for repeated

offend-ers (usoffend-ers who distribute MDI) in the news feed

Decrease in visibility of MDI, hence the user’s interaction with it

Content-focused +

Context-focused Facebook, YouTube, Reddit, Twitter 4 Deleting accounts of MDI

crea-tors and bots Decreases the quantity of MDI Content-focused Facebook, YouTube, Twitter 5 Adding a visible link to official

information on COVID-19 from WHO

Increases visibility of accurate information, presumably edu-cates the users

Content-focused +

Context-focused Facebook, YouTube, Twitter

6 Creating a dedicated page for local news and official informa-tion on COVID-19, linking to it visibly

Increases visibility of accurate information, presumably edu-cates the users

Content-focused +

Context-focused Facebook, Twitter

7 Displaying “related” items below posts about COVID-19, most of which are leading to fact-checked information

Increases diversity of the infor-mation to which the user is exposed, presumably educates the users

Context-focused +

(7)

L. Marin

1 3

it or even share it. Most of these emotional reactions have dedicated buttons which can be clicked mindlessly, yield-ing the interaction seamless. In the wake of the Covid-19 pandemic, Facebook even added a new emoticon expressing a reaction of solidarity (see Fig. 1). The assumption was that now, more than ever, users needed to express emotions online with a richer palette. However the simplistic way of expressing such emotions did not change, it was part of the interaction design from the beginning.

The emotional infrastructure of social media was not something requested by users but designed from the start. Major social media platforms are oriented towards maximis-ing the user’s engagement, i.e. how much time one spends on a specific platform and how much attention is consumed (Whiting and Williams 2013). These kinds of interactions actively promote the user’s “attention bulimia” i.e. a behav-iour oriented towards “maximizing the number of likes” (Del Vicario et al. 2016, p. 1) and presumably other posi-tive reactions. Most buttons for emotional reactions are of positive emotions (like, love, hug, laugh) while in recent years Facebook added some more nuanced emotions such as angry, sad or cry. But the overwhelming effect of these emotional reactions is to make other users feel liked by their social network hence, to make the platform a place where one wants to keep returning to for emotional gratification.

For many users confined to their homes by the pandemic, social media platforms became a window to the world, as the television set used to be in older days and the easiest way of relating to others. In such times of distress and uncertainty, users posted more frequently than usual (Cinelli et al. 2020) but some of the information posted was not meant to inform others, but rather to express one’s concerns and emotions related to the pandemic. Posts were also meant to get reac-tions from one’s friends in an attempt to confirm that the others were also feeling the same way as one does. Post-ing about the pandemic became a strategic way of gaugPost-ing other’s emotions on the crisis situation and gathering some

feeling of consensus from one’s social network. The consen-sus sought on social media was of an emotional nature which may be at odds with an epistemic consensus about the nature of the facts at hand.

The weak epistemic context

During the pandemic, several epistemologists and philoso-phers of science stepped up and tried to educate the general public on what sources to trust as experts and how to discern facts from fiction about the pandemic—in podcasts, opinion pieces and on their social media accounts (Weinberg 2020). While this effort is laudable, it needs to be complemented with another approach, taking into account the wider epis-temic context in which information travels on social media. This is a particularly weak epistemic context in which infor-mation is not always shared to inform. Social media plat-forms are not places where one usually goes to be informed. At least in regular, day to day situations, users turn to social media platforms to relate, to communicate and to be enter-tained (Fuchs 2014). The weak epistemic context of social media is ruled by serendipity (Reviglio 2017), meaning that many users get to be informed by accident.

In a crisis situation, users tend to change how they use the platform and shifting towards the communication of vital information such as imminent risks or their location and also seeking to be informed by latest developments from people from the local site of the disaster. The entertainment func-tion tends to become secondary in emergencies (Zeng et al. 2016). In the 2020 pandemic situation, the difference was that the crisis was global and that the duration was rather long. This time, the uncertainty that accompanies a crisis situation was extended over months. As epistemic agents, online users tried to make sense of what was going on with them, what they could expect and to assess the personal risks, over a longer period.

(8)

Three contextual dimensions of information on social media: lessons learned from the COVID-19…

1 3

The pandemic was an extended crisis situation com-pounded with social alienation on top. This made users feel lost and overwhelmed by problems one could not under-stand. Hence the desire—legitimate to a point—for everyone to be an expert so that they could at least understand what was happening to them. People did not want to be experts in epidemiology, quarantine measures, and home remedies for viruses because of a sudden surge of intellectual curiosity. They needed a way of coping that was also understandable to them. Meanwhile, the official discourse of “trust the experts” and “please don’t share information you do not understand” incapacitated them as epistemic agents. Requiring users to do nothing and just comply went against the general desire to do something, as a way to take control. Given the increase in posts on the pandemic by regular users, it may seem that many have tried to become experts overnight in epidemiol-ogy, viruses and vaccines. The comic below (Fig. 2) illus-trates the frequent situation emerging during the pandemic of members of the lay public hijacking the role of the expert.

A discussion on conditions of trust and expertise makes sense in a regular epistemic context when agents try to acquire knowledge about a domain they know nothing about, having to choose which experts to trust (Goldman and O’Connor 2019). However, acquiring knowledge was probably not the main goal of social media users who started posting scientific information which they did not understand. Rather, many users tried to build some understanding of the situation, to make sense of the events. In these cases, shared understanding in a circle or network of friends seemed to be more important for users than gaining access to exper-tise. The scientific information was posted by lay users to back one’s personal opinions, to urge for a certain course

of action, or to gather consensus. Given these weak epis-temic uses of information targeted at emotional fulfilment and networking purposes, the regular content-focused meas-ures would probably be less effective than predicted on such users.

The strong normative context

Related to the previous point and stemming from it, most factual information shared around on social media had some normative implications which often shadowed any knowl-edge claim. Descriptive information was used for prescrip-tive or evaluaprescrip-tive aims. Scientific expertise was co-opted strategically to enforce one’s own pre-existing evaluative opinions. Typical MDI claims are not merely descriptive claims of a state of affairs in the world, but often embedded in a normative context be those prescriptive or evaluative claims, both types are meant to change attitudes of the online users. MDI was shared because it prescribed actions or led to evaluations of the state of affairs which users already agreed with. Hence debunking the facts would have solved only half of the puzzle, since the user’s motivation to believe these normative claims would have not been dealt with.

One example of the strong normative context for MDI, also involving a clear “politicisation” of MDI during the pandemic (Howard 2020), concerns one of the most popular types of claims analysed by EU vs Disinfo (2020) in which the EU was depicted as powerless and scattered in dealing with the pandemic. This claim was traced back to Russian-backed agencies which aimed to make users believe that, ultimately, Russia was stronger than the EU (Howard 2020). Such claims can be debunked by showing that there were

(9)

L. Marin

1 3

coordinated measures taken by the EU, however, the implicit claim that other states dealt better with the pandemic than the EU is hard to debunk since it is not explicitly stated. This is just one type of difficulty with MDI which cannot be tackled with a content-focused approach: implicit evaluative claims in which one term of the comparison is not named.

Some evaluative claims can be checked (if these involve relational predicates which are measurable such as “better than” or “more efficient than”) however others, incidentally the politicised evaluative claims, are harder to assess. In Russian-backed claims against the EU, the name of “Rus-sia” is not mentioned anywhere in the text of the “news”, since the aim is to erode the trust in EU from its citizens. If these citizens happen to be in Eastern Europe, this erosion of trust could lead to an anti-EU generalised feeling, and ultimately bottom-up pressures to exit the EU. These kinds of campaigns cannot be easily fact-checked since the effect is achieved by playing the long game. What looks like news about the pandemic is a dog-whistle about something else.

The strong normative context is visible also when using scientific expertise co-opted to back up prescriptive claims otherwise untenable. One example is an unpublished paper by Blocken et al. titled “Towards aerodynamically equivalent COVID19 1.5 m social distancing for walking and running” (2020) in which an animated image showed a simulation of how joggers coughing will spread particles of droplets when running at a distance of 1.5 m from each other. The paper became viral on social media despite not being peer-reviewed nor published on a pre-print website. The visual animation showing the spread of droplets was understand-able by every lay member of the public, without needing to have specialised knowledge of actual aerodynamics, and presumably made the paper so popular among non-scientists who used it to make prescriptive claims by non-scientists. While the authors hypothesised that it might be unsafe to run close to another—and that even 1.5 m distance might not be enough for jogging—the social media audience took this as a reason to shame the runners in their neighbourhoods (Koe-bler 2020). Meanwhile, the first author of the study posted a document on his website answering certain questions about the study3 and refused to draw any epidemiological

conclu-sions, urging for other’s expertise. But social media users did not shy from becoming experts and drawing the conclusions themselves, as the information in the Blocken et al. paper was just ammunition in a larger informational battle about what others should do.

Even if the scientific claims of regular users are checked, their aim remains to prescribe actions for others and to eval-uate the world in a way that will be endorsed by one’s com-munity of friends. For these purposes, other pieces of news

will be co-opted if the first ones were flagged as hoaxes. Content-based approaches are then ineffective against this strong desire of social media users to emit evaluative or pre-scriptive claims about the world and strategically use sci-ence-looking sources to back these up. One should address the very desire of regular users to evaluate the world from the little soap-boxes that social media affords.

Avenues for future development: contextual

approaches to informational disorders

online in times of crises

The three dimensions of the informational context on social media previously mentioned (strong emotional, strong nor-mative and weak epistemic) have been analytically distin-guished but they function simultaneously to promote certain user behaviours which one could call irresponsible informa-tion sharing on social media. The weak epistemic context re-enforces the strong normative claims which are coupled with emotional reactions leading users to become strongly attached to their claims and indifferent to their debunking. While the three dimensions of the informational context work together to produce a perfect storm of low-quality and redundant information—an infodemic—one could still try to design modes of interaction to deal with each of them.

Future measures dealing with MDI and infodemic in a global pandemic need to also tackle the high emotional load of most MDI items. A practical way of flagging this could be devised by modelling on Facebook’s “hoax alert” system which warns users that a certain post they are reading has been fact-checked and is probably a hoax. A similar system could be designing an “emotional alert” flag which appears below certain news-looking items which have an unusually high amount of emotional triggering words. This kind of alert would show to users when certain news-like items lead them to feel something quite specific. The readers would be still free to engage with such items, but at least they would be warned about possible emotional manipulation. Such alert would also flag click-bait and irrelevant news which are not false in themselves, but which do contribute to the infodemic because of the high virality and potential for polarisation.

The weak epistemic context needs to be tackled together with the high normative one. When users engage with news-items relating to the pandemic (or any other crisis situation), they can be shown alternative news and pages or users to follow which are experts in the field. This has been already implemented, but with little success as it needs to be com-plemented by more context-aware measures. A new measure could be to try to increase the users’ critical engagement with a certain class of news items. This could happen by posting a small survey under the tricky news items asking the users to answer the following questions: “What does

(10)

Three contextual dimensions of information on social media: lessons learned from the COVID-19…

1 3

this news lead me to believe?”, “What does this post ask me to feel?”, “Please rate how strongly you agreed with these claims before reading this post”. And, finally, after answer-ing this mini-survey, the user could be shown a disclaimer stating, “Now that you’ve read posts you agreed with, how about trying something different?”—and the alternative sources with diverse information from experts could be dis-played, but after the user has been primed to be more critical and diversify one’s information sources. To incentivise users to fill in these surveys, they could receive certain bonus points on the social media platform and, once they accumu-late a certain number of points, they could get a badge next to their name of “critical media user” or “critical thinker” which would ensure visibility of their informational skills.

These sample measures I have proposed—and presum-ably other context-sensitive measures—are fit to be imple-mented only in crisis situations and only targeting the user’s posts about the current crisis situation. It is, of course, pos-sible to integrate these measures in the day to day users’ interaction. However, social media fulfils certain deep emo-tional needs of users such as expressing strong normative opinions and seeking emotional consensus. If these contexts are strongly discouraged, the users may find other platforms to do so and they may abandon the too critical platforms. Meanwhile, in a crisis situation, these dimensions could be targeted specifically and confined to users that share MDI and redundant information about the crisis at hand.

This paper outlined three types of informational context which have not yet been taken into account by the current measures to curb down disinformation and the amount of irrelevant information on social media platforms. MDI countermeasures taken thus far have a limited effect because these did not account for the diversity of information con-texts on social media. These measures do not address the pri-mary causes of the infodemic itself, namely the user’s need to post content as a way of making sense of the situation and of expressing one’s opinion as a way of gathering con-sensus reactions. In light of these considerations, we need a value-sensitive design re-assessment of how users inter-act on social media among themselves as well as with the information found online in crisis situations. Designing for thoughtful, critical and meaningful user interaction should become an explicit aim for the future of social media plat-forms in times of pandemics and other global emergencies.

Acknowledgements The author would like to thank Samantha Marie

Copeland for commenting on an earlier draft of this article and to the two reviewers for their insightful comments on the first version of the article.

Funding This project has received funding from the European Union’s

Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie Grant Agreement Number. 707404. The opin-ions expressed in this document reflect only the author’s view. The

European Commission is not responsible for any use that may be made of the information it contains.

Compliance with ethical standards

Conflict of interest No conflicts of interest.

Open Access This article is licensed under a Creative Commons

Attri-bution 4.0 International License, which permits use, sharing, adapta-tion, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creat iveco mmons .org/licen ses/by/4.0/.

References

Bakir, V., & McStay, A. (2018). Fake news and the economy of emotions. Digital Journalism, 6, 154–175. https ://doi. org/10.1080/21670 811.2017.13456 45.

Blocken, B., Malizia, F., van Druenen, T., & Marchal, T. (2020). Towards aerodynamically equivalent COVID19 1.5 m social dis-tancing for walking and running. Questions and Answers. Website Bert Blocken, Eindhoven University of Technology (The Nether-lands) and KU Leuven (Belgium). Disponibile su: http://www. urban physi cs.net/COVID 19.html (ultimo accesso 21 aprile 2020). Brennen, J. S., Simon, F., Howard, P. N., & Nielsen, R. K. (2020).

Types, sources, and claims of COVID-19 misinformation. Reuters Institute. Retrieved from https ://reute rsins titut e.polit ics.ox.ac.uk/ types -sourc es-and-claim s-covid -19-misin forma tion. Accessed 17 Apr 2020.

Cinelli, M., Quattrociocchi, W., Galeazzi, A., Valensise, C. M., Brugnoli, E., Schmidt, A. L. et al. (2020). The COVID-19 social media infodemic. https ://arxiv .org/pdf/2003.05004 v1.

Del Vicario, M., Vivaldo, G., Bessi, A., Zollo, F., Scala, A., Cal-darelli, G., et al. (2016). Echo chambers: Emotional contagion and group polarization on facebook. Scientific Reports. https :// doi.org/10.1038/srep3 7825.

EU vs Disinfo. (2020). Throwing coronavirus disinfo at the wall to see what sticks. EU vs Disinfo. Retrieved from https ://euvsd isinf o.eu/throw ing-coron aviru s-disin fo-at-the-wall-to-see-what-stick s/. Accessed 20 Apr 2020.

Fallis, D. (2014). The varieties of disinformation. In L. Floridi & P. Illari (Eds.), The philosophy of information quality (pp. 135–162). Cham: Springer International Publishing.

Floridi, L., & Illari, P. (Eds.). (2014). The philosophy of information quality. Cham: Springer International Publishing.

Friedman, B. (1996). Value-sensitive design. Interactions, 3(6), 16–23. Fuchs, C. (2014). Social media: A critical introduction. London:

SAGE.

Goldman, A., & O’Connor, C. (2019). Social epistemology. In: E N. Zalta (Ed.), The Stanford encyclopedia of philosophy (2019th ed.): Metaphysics Research Lab, Stanford University.

Howard, P. (2020). Misinformation and the coronavirus resistance. Oxford Internet Institute. Retrieved from https ://www.oii.ox.ac.uk/blog/misin forma tion-and-the-coron aviru s-resis tance /. Accessed 1 May 2020.

(11)

L. Marin

1 3

Khaldarova, I., & Pantti, M. (2016). Fake news. Journalism Practice, 10, 891–901. https ://doi.org/10.1080/17512 786.2016.11632 37. Koebler, J. (2020). The viral ‘Study’ about runners spreading

coronavi-rus is not actually a study. Vice. Retrieved from https ://www.vice. com/en_us/artic le/v74az 9/the-viral -study -about -runne rs-sprea ding-coron aviru s-is-not-actua lly-a-study .

Koroleva, K., Krasnova, H., & Günther, O. (2010). Stop spamming me! Exploring information overload on Facebook. In: Proc 16th AMCIS, Lima.

Naeem, S. B., & Bhatti, R. (2020). The Covid-19 ’infodemic’: A new front for information professionals. Health Information and Libraries Journal. https ://doi.org/10.1111/hir.12311 .

Reviglio, U. (2017, November). Serendipity by design? How to turn from diversity exposure to diversity experience to face filter bub-bles in social media. In International Conference on Internet Sci-ence (pp. 281–300). Springer, Cham.

Schilit, B., Adams, N., & Want, R. (1994). Context-aware comput-ing applications. In 1994 First Workshop on Mobile Computcomput-ing Systems and Applications (WMCSA), Santa Cruz, California, USA, 08.12.1994 - 09.12.1994 (pp. 85–90): IEEE. https ://doi. org/10.1109/WMCSA .1994.16.

Silverman, H. (2019). The next phase in fighting misinformation. Face-book Newsroom.

Tangcharoensathien, V., Calleja, N., Nguyen, T., Purnat, T., D’Agostino, M., Garcia-Saiso, S., et al. (2020). Framework for managing the COVID-19 infodemic: Methods and results of an online, crowdsourced WHO Technical Consultation. Journal of Medical Internet Research, 22, e19659. https ://doi. org/10.2196/19659 .

Van den Hoven, J. (2007). ICT and value sensitive design. The infor-mation society: Innovation, legitimacy, ethics and democracy in honor of Professor Jacques Berleur SJ (pp. 67–72). Boston: Springer.

Wardle, C., & Derakhshan, H. (2018). Thinking about ‘information disorder’: Formats of misinformation, disinformation, and mal-information. In: C. Ireton & J. Posetti (Eds.), Journalism, “fake news” et disinformation: Handbook for journalism education and training (pp. 43–54, UNESCO series on journalism educa-tion). Paris: United Nations Educational, Scientific and Cultural Organization.

Weinberg, J. (2020). The role of philosophy & philosophers in the coronavirus pandemic. Daily Nous. Retrieved from https ://daily nous.com/2020/03/27/role-philo sophy -philo sophe rs-coron aviru s-pande mic/. Accessed 15 June 2020.

Whiting, A., & Williams, D. (2013). Why people use social media: A uses and gratifications approach. Qualitative Market Research: An International Journal, 16, 362.

Wixon, D., Holtzblatt, K., & Knox, S. (1990). Contextual design: An emergent view of system design. In (pp. 329–336, CHI ’90). New York, NY, USA: Association for Computing Machinery. https :// doi.org/10.1145/97243 .97304 .

World Health Organization. (2020). Novel Coronavirus(2019-nCoV): Situation Report-13. WHO. Retrieved from https ://www.who.int/ docs/defau lt-sourc e/coron aviru se/situa tion-repor ts/20200 202-sitre p-13-ncov-v3.pdf. Accessed 30 Mar 2020.

Zeng, L., Starbird, K., & Spiro, E. S. (2016). Rumors at the speed of light? Modeling the rate of rumor transmission during crisis. In (pp. 1969–1978).

Zollo, F., Novak, P. K., Del Vicario, M., Bessi, A., Mozetič, I., Scala, A., et al. (2015). Emotional dynamics in the age of misinforma-tion. PLoS ONE, 10, e0138740. https ://doi.org/10.1371/journ al.pone.01387 40.

Publisher’s Note Springer Nature remains neutral with regard to

Cytaty

Powiązane dokumenty