Aljaž Zrnec, Marko Poženel and Dejan Lavbič. 2022. Users’ ability to perceive misinformation: An information quality assessment approach, Information Processing & Management (IPM), 59(1).

Abstract

Digital information exchange enables quick creation and sharing of information and thus changes existing habits. Social media is becoming the main source of news for end-users replacing traditional media. This also enables the proliferation of fake news, which misinforms readers and is used to serve the interests of the creators. As a result, automated fake news detection systems are attracting attention. However, automatic fake news detection presents a major challenge; content evaluation is increasingly becoming the responsibility of the end-user. Thus, in the present study we used information quality (IQ) as an instrument to investigate how users can detect fake news. Specifically, we examined how users perceive fake news in the form of shorter paragraphs on individual IQ dimensions. We also investigated which user characteristics might affect fake news detection. We performed an empirical study with 1123 users, who evaluated randomly generated stories with statements of various level of correctness by individual IQ dimensions. The results reveal that IQ can be used as a tool for fake news detection. Our findings show that (1) domain knowledge has a positive impact on fake news detection; (2) education in combination with domain knowledge improves fake news detection; and (3) personality trait conscientiousness contributes significantly to fake news detection in all dimensions.

Keywords

Fake news, Misinformation, Information quality, Big Five personality traits, Fake news detection, Case study

1 Introduction

The rapid integration of digitalisation in our lives increases the volume of data in the global data sphere and changes users’ behaviour. The availability of various multimedia devices, constant connection to the internet, and overflowing information sources affect the way people access existing media (Chang and Hwang 2020). Traditional media are facing challenges accessing users in this new reality. Users are moving away from traditional to social media when seeking information. The reasons for this change are the possibilities that online information access offers, including timeliness, ease of access and an endless supply of knowledge and information. Undoubtedly, social media has significantly contributed to this transformation (Bondielli and Marcelloni 2019). People can follow official social media channels and post on their networks, while journalists may use social media to gain public opinions on certain current topics and even discover potential new stories (Bondielli and Marcelloni 2019). Allcott and Gentzkow (2017) state that \(62 \%\) of US adults obtain news on social media. Also the case study for Britain shows the growing importance of social media and the increasing share of social media news consumption (Newman, Dutton, and Blank 2013).

The way news spreads on social media is different from its spread via former media technologies. Content may reach users without prior filtering, fact-checking, or editorial judgement, while news can be created online faster and cheaper. Unfortunately, that makes social media a possible abundant source of unverified and/or false information. The attractiveness of spreading misinformation lies in the fact that Websites and individuals may profit from spreading misinformation. For example, teenagers from Macedonia became wealthy by penny-per-click advertising during U.S. presidential elections (Zhou and Zafarani 2020). False news is seen as a great threat to public trust, the economy, and journalism. The problem of misinformation detection is a challenging task since misinformation appear in different forms and is made seemingly real.

In the literature, many approaches for misinformation detection have been proposed (Bondielli and Marcelloni 2019; Zhou and Zafarani 2020). Due to the rapid diffusion of misinformation in social media and limited scalability of manual detection methods, automatic fact-checking methods gained the attention of the researchers (Shu et al. 2017; Gravanis et al. 2019). However, the automatic detection of fake news based on content is non-trivial. Solutions are usually specialized, and more importantly, do not involve comprehensive human knowledge of the domain. Besides regular news, a large body of fake news can also be created and disseminated for a variety of purposes, such as financial or political gain (Shu et al. 2017), and can have significant effects on society and public opinion (X. Zhang and Ghorbani 2020).

The terms false information and fake news may get confused with various other terms associated with inaccurate information. However, for online content, the term “fake news” has been de-facto used to denote false information in mainstream media (Bondielli and Marcelloni 2019). Existing studies often use the term fake news with related concepts. Wu et al. (2019) defined misinformation as an umbrella term to include all false or inaccurate information that is spread in social media. They organize different type of information into these key categories: misinformation (Kucharski 2016; Wardle 2017), disinformation (Kshetri and Voas 2017; Wardle 2017), fake news (Shu et al. 2017), rumour (Zubiaga et al. 2018), urban legend, spam and troll. These terms all share a characteristic that the inaccurate messages can cause distress and various kinds of destructive effect through social media, especially when timely intervention is absent (Wu et al. 2019). They also introduce the concept of unverified information, which can be false or true and accurate. Zhou and Zafarani (2020) also presented similar categorization of concepts that frequently co-occur or have overlapped with fake news: deceptive news (Allcott and Gentzkow 2017; Lazer et al. 2018; Shu et al. 2017), false news (Vosoughi, Roy, and Aral 2018), satire news (Conroy, Rubin, and Chen 2015; C. Tandoc Jr., Lim, and Ling 2018; Wardle 2017), disinformation, misinformation, cherry-picking (Asudeh et al. 2020), clickbait (Chen, Conroy, and Rubin 2015), and rumour (Zubiaga et al. 2018).

Misinformation and disinformation both refer to fake or inaccurate information. However, disinformation is deliberately created to mislead the user while misinformation is not (Wu et al. 2019). Despite the popularity of fake news, there has been no universal definition for the term. In literature, Wu et al. (2019) define fake news as intentionally spread misinformation that is in the format of news. Zhou and Zafarani (2020) provide a broad definition of fake news as false news, and a narrow definition as intentionally false news published by a news outlet. In this paper, we will adopt the broad definition of fake news presented in Zhou and Zafarani (2020). Our approach focuses on the ability to detect fake news for a regular user, based on information quality that the source contains.

Existing studies that explore fake news on social media mainly focus on its automatic detection. The rising quantity of fake news on social media and its rapid dissemination, make automatic detection approaches attractive (Bindu, Mishra, and Thilagam 2018). Relatively few studies have investigated fake news detection from the user’s (news consumer’s) perspective. In social media, fake news can take several forms and shapes and is becoming more and more similar to proper news, making it more difficult to efficiently detect, both manually and automatically (Bondielli and Marcelloni 2019). With the growing exposure of end-users to unverified information, users become indirectly responsible for content assessment. From this observation, we argue that it is worthwhile to investigate how successful users are in detecting fake news.

Our approach relies on an individual acting as a self fact-checker. Compared to expert-based fact-checking, an individual has less knowledge of a specific topic or domain and leads to less accurate results. It also poorly scales with volume due to the fatigue as a result of effort someone has to invest to critically evaluate contents. However, cautious users can employ the services of expert-based or crowd-sourced fact-checking websites (like Politifact, or Fiskitt respectively) and verify ground truth for the detection of fake news in areas where they do not feel confident.

For this purpose, we conducted a study where participants identified fake news using an instrument of information quality (IQ). R. Y. Wang and Strong (1996) define IQ as information that is suitable for information users to make use of and Gustavsson and Wänström (2009) as the ability to satisfy the stated and implied needs of the information consumer. It is a multidimensional construct, consisting of several dimensions that describe certain aspects of IQ and can be measured and assessed (Gustavsson and Wänström 2009; Paggi et al. 2021). We first build upon existing studies on assessing source documents using IQ as an instrument (Yaari, Baruchson-Arbib, and Bar-Ilan 2011). Second, we argue that fake news affects IQ and can be detected. Previous research has studied agreement between assessors when assessing IQ based on IQ dimensions and thus shown that IQ can be perceived and measured (Arazy and Kopak 2011; Fidler and Lavbič 2017). Following their findings, fake news might reflect on IQ dimensions, and it might be possible to detect them by the assessor.

Assessing IQ has proved to be difficult (Arazy, Kopak, and Hadar 2017). Hence, fake news detection based on IQ might also be a difficult task. In this study, we sought to investigate how successfully users can detect fake news without the help of external tools and applications. Users assess the information that they are familiar with, with higher confidence than the information that they are unfamiliar with. Thus, in our study, we controlled for the influence of foreknowledge, because we argue that it is one of the vital determinants of fake news detection.

Existing studies have also investigated the positive correlation of formal education with observed variables (e.g. the impact of formal education on lodging professionals’ management success (Solnet et al. 2010) and the degree of payment inclusion (Polasik, Huterska, and Meler 2018)). We assume that formal education might also impact fake news detection score. Thus we also examined the impact of formal education on the ability to detect fake news.

Besides users’ core ability to detect fake news and familiarity with the source topics, personality traits (Gosling, Rentfrow, and Swann Jr 2003) might also affect users’ ability to detect fake news. Previous studies have investigated the relationship between personality traits and online content (Mehdizadeh 2010; Amichai-Hamburger and Vinitzky 2010; Ehrenberg et al. 2008; Wilson, Fornasier, and White 2010; Correa, Hinsley, and De Zuniga 2010). However, their findings may not apply to fake news detection abilities. The question arises as to whether and in what way personality traits affect the ability to detect fake news. Thus, we sought to expand the literature by investigating the relationship between personality traits and the ability to detect fake news. Our study is based on the widely accepted psychological Big Five personality trait model, which measures individuals’ personality traits (Correa, Hinsley, and De Zuniga 2010) and provides a solid personality framework (Yin, Zhang, and Liu 2020).

This paper contributes by examining whether end-users can identify fake news using IQ as an instrument. To the best of our knowledge, no other studies investigate the use of IQ instrument for fake news detection. Since little is known about fake news detection using IQ, we seek to focus the research community’s attention in this area. By investigating the impact of fake news on individual IQ dimensions, this study provides steps for fake news detection and confirms that fake news can be identified on the basis of information distortion. Further, our study examines the impact of education, domain knowledge and personality traits on fake news detection score.

The remainder of the paper is organised as follows. In section 2 we provide a review of related work regarding fake news detection, IQ instrument, and Big Five personality trait framework. In section 3, we introduce the problem statement and the research questions. In section 4 the process of conducting the experiment and data collection is outlined, while in section 5, empirical results, discussion, practical implications, and limitations are presented. Finally, in section 6, we present conclusions, and suggestions for future work.

3 Problem statement

As the volume of content grows, it is more and more difficult to acquire complete and accurate information. We argue that the approach of recognizing fake news based on IQ is worth exploring. To identify fake news based on the selected set of dimensions, we analysed short news of differing IQ quality and studied how individual IQ dimensions correlated with fake news. In this work, we highlight the relationship between IQ and susceptibility to fake news.

Only a few researchers have employed IQ in studies of fake news. Agarwal and Yiliyasi (2010) investigated the relationship between \(16\) IQ dimensions and different types of social media. However, that study did not focus on how fake news affects IQ dimensions. Zhu, Li, He, et al. (2020) analysed the differences in users’ perceived quality of online reviews on \(10\) diverse IQ dimensions, while a study from Zhu, Li, Wang, et al. (2020) examined the effect of perceived IQ on purchase intention. Neither of these studies focused on fake news. The aforementioned studies also used an extended range of IQ dimensions. Some previous studies that examined the impact of content on IQ dimensions (Arazy and Kopak 2011; Arazy, Kopak, and Hadar 2017) used a reduced set of dimensions (i.e. accuracy, objectivity, representation and completeness) rather than attempting to cover the extended range. Like Arazy, Kopak, and Hadar (2017), we believe that this subset of dimensions reasonably represents different aspects of IQ. Furthermore, our use of a reduced set makes the study more directly comparable to previous work in the field.

People access online information from various sources that vary in terms of functionality and content length. We believe that content’s length has an important impact on IQ and should not be neglected. In terms of functionality, Agarwal and Yiliyasi (2010) proposed the following categorization of social media sites, which are arranged in a descending length order: wikis, social news, blogs, social networks, microblogging (tumblelogs), media sharing, and bookmarking. Wikis are publicly edited encyclopaedias and are usually longer texts. In general, blogs can be long, but usually contain fewer than \(2,000\) words in an individual blog post. Text posts on social networks (e.g. Facebook) are much shorter than that. The optimal post length on social media is between \(40\) and \(60\) characters, while microblogs (e.g. tweets) are further limited to \(160\) characters. Lahuerta-Otero and Cordero-Gutiérrez (2016) investigated the characteristics of influencers tweets and found that tweet length does not have a significant effect on microblogging users’ influence. Wikipedia has been used several times as a source in previous IQ research (Luyt and Tan 2010; Fallis 2008; Lim 2009; Arazy and Kopak 2011; Arazy, Kopak, and Hadar 2017). A limited number of papers have investigated other categories from an IQ perspective. Zhu, Li, He, et al. (2020) and Zhu, Li, Wang, et al. (2020) studied online reviews. A typical online review length is from \(100\) to \(1,000\) words (Iio 2012), so we categorize these as blogs. In their studies, Mai (2013) and Fidler and Lavbič (2017) focused on documents that are shorter than wikis and typical blogs. They used a cooperative principle (Mai 2013) to shorten original Wikipedia articles to make them more manageable for typical Wikipedia users (e.g. students) while still retaining source IQ. In this study, we focused on an even smaller unit of text – individual paragraphs. The paragraphs under study consist of a few sentences that refer to various IQ dimensions. We sought to investigate how users can identify fake news when dealing with paragraph-size source texts. The paragraphs are longer than posts on social networks and shorter than the documents studied by Fidler and Lavbič (2017).

Several automatic approaches prevent false news from reaching end-users (Wu et al. 2019; Zhou and Zafarani 2020; Viviani and Pasi 2017). While these approaches have several advantages, they do not necessarily isolate users from fake news. We believe that end-user detection capabilities should also be addressed and investigated. Pennycook and Rand (2020) supported this idea – claiming that it is important to understand the cognitive factors that allow readers to weed out the untrue in favour of the true. Kim and Dennis (2018) also stated that media has moved quality control for detecting fake news from trained journalists to regular users. Based on the idea that research should highlight the side of the information consumer, we investigated how successfully users can detect fake news based on the IQ carried by source documents.

Fake news detection using IQ is closely related to the end-user’s (i.e. news consumer) perception of the content. Previous studies have investigated how personality traits influence the perception of content. For example, Pennycook and Rand (2020) investigated the psychological profile of individuals who fall prey to fake news. They suggested that there is a connection between the personal traits of users and receptiveness of users for pseudo-profound constructed fake news for garnering attention. The authors do not highlight the problem from an IQ point of view. Thus, our research aims to explore the impact of personality traits on the detection of fake news based on IQ dimensions.

In this study, we focused on determining whether it is possible to infer that some information is false based on information distortion. The same concept within the topic \(t\) was in our experiment described from the aspect of selected IQ dimension \(S\). A total of \(24\) statements per given topic \(t\) for each of the \(4\) IQ dimensions (accuracy, objectivity, representation and completeness) were constructed, resulting in half of the statements with a negative level of correctness \(l^{-}\) and the other half with a positive level of correctness \(l^{+}\) as further discussed in section 4. We conjectured that fake news is likely to reflect individual dimensions of IQ. We tried to detect fake news based on the quality of the information the media contained by user evaluation \(IQ_{init}(t,S,i)\) of a given statement from topic \(t\), from an IQ dimension \(S\) and in \(i\)-th story. To summarise, the input to our experiment was a randomly generated story, constructed of 4 short statements, each related to an IQ dimension \(S\) with a varying level of correctness \(l\). The output was the user’s perception of IQ for a given story \(i\), i.e. \(IQ_{init}(t,S,i)\). We stress that, while the literature on IQ focuses mainly on IQ value and its measurability, in this work, we investigated the lack of IQ as an indicator of misinformation.

Past research has shown that some IQ dimensions are easier to assess than others (Arazy and Kopak 2011; Arazy, Kopak, and Hadar 2017). We sought to investigate if that also holds for fake news identification based on IQ dimensions. Furthermore, we also investigated which cues (e.g. foreknowledge, education) affect fake news detection. We aimed to further explore the correlation between evaluators’ personality traits and the perception of news as fake or genuine. In general, we studied whether IQ can successfully be used as a tool for fake-news detection.

Fake news detection using IQ is closely related to the end-user’s (i.e. news consumer) perception of the content. Previous studies have investigated how personality traits influence the perception of content. Pennycook and Rand (2020) suggested that there is a connection between the personal traits of users and receptiveness of users for pseudo-profound constructed fake news for garnering attention. The authors do not highlight the problem from an IQ point of view. In this study, we investigate how personality traits affect fake news detection using the Big Five framework. We examine how personality traits represented by the Big Five personality dimensions affect fake news detection in terms of individual IQ dimensions. We measure personality traits using a custom, brief instrument based on a ten-item personality inventory (TIPI) (Gosling, Rentfrow, and Swann Jr 2003).

To the best of our knowledge, no prior study has empirically investigated IQ as a tool for fake news detection. We designed an empirical study to examine the fake news detection ability of different users and formulated the following research questions (RQ):

  • RQ1: Are users able to perceive the IQ of short news and identify fake news?
  • RQ2: Which IQ dimensions make users more likely to detect fake news?
  • RQ3: Does education or domain knowledge affect users’ fake news detection?
  • RQ4: Which personal traits facilitate fake news detection and which IQ dimension do these traits affect the most?

4 Experiment

An online web application in a form of an interactive questionnaire was developed to explicitly support all steps of the experiment (see Figure 4.1) and the link was distributed to \(\boldsymbol{1123}\) participants (\(\boldsymbol{31.6\%}\) female, \(\boldsymbol{68.4\%}\) male) from \(\boldsymbol{15}\) to \(\boldsymbol{69}\) years of age, with a median age of \(\boldsymbol{20}\). We targeted a population with various levels of education by directly inviting students from High School and University. To increase the diversity of participants and enhance the external validity of the research, we also reach out to other participants. The distribution of our sample was \(\boldsymbol{57.4\%}\) undergraduate students and \(\boldsymbol{42.6\%}\) non-students. The study focuses on the European region, where the participation in the study was voluntary with no paid compensation.

The questionnaire was divided into an introductory portion, with the collection of demographic information about participants, and the main portion that asked participants to evaluate several statements with varying level of correctness (inverse to a level of fakeness) on selected topics in terms of four IQ dimensions: objectivity, representation, accuracy, and completeness (Table 4.1). A similar study using a comparable approach to design the questionnaire and conduct the experiments was used in Fidler and Lavbič (2017), Lee et al. (2002) and Yaari, Baruchson-Arbib, and Bar-Ilan (2011).

Table 4.1: Selected IQ dimensions adopted from related work and used in this study
Dimension Description
Accuracy Information in the statement is accurate.
Completeness Statement is complete and includes all necessary information.
Objectivity Statement is objective; it represents objective opinion about presented topic.
Representation Statement is presented consistently and formatted concisely.

The motivation for conducting the research was presented to participants in the introductory portion of the questionnaire; they were then asked to provide some general information and their familiarity (knowledge level) with the topic of the statements being evaluated. Participants also carried out a self-assessment of their personality traits based on the Big Five personality model (Lamers et al. 2012).

Several measuring instruments have been developed to measure the Big Five dimensions. In general, they are too lengthy for scenarios where participants’ time is important (Woods and Hampson 2005). Ehrhart et al. (2009) state that in such circumstances, an extremely short measure of the Big Five is adequate, and that researchers should consider the TIPI instrument (Gosling, Rentfrow, and Swann Jr 2003). We retained the same ten items from the TIPI instrument for measuring Big Five personality dimensions and asked participants to identify themselves in terms of positive and negative personality traits.

The average participant in our study was an individual with positive extraversion, neutral neuroticism, positive conscientiousness, positive agreeableness and positive openness to experience. The distribution of participants in Big5 groups was nearly uniform for Big5 factors neuroticism (\(35.3\%\) positive, \(35.8\%\) neutral and \(28.9\%\) negative), extraversion (\(41.3\%\) positive, \(31.7\%\) neutral and \(27\%\) negative) and conscientiousness (\(43.3\%\) positive, \(37.7\%\) neutral and \(19.1\%\) negative). Big5 factors agreeableness (\(57.3\%\) positive, \(33.2\%\) neutral and \(9.5\%\) negative) and openness to experience (\(49.1\%\) positive, \(41.9\%\) neutral and \(9\%\) negative) had a higher share of individuals within a positive Big5 group.

Algorithm 1 and Figure 4.1 illustrate all steps in the experiment process.

Algorithm 1: Steps in the experiment process
set topics T = { "the Occitan Language", "Alexandrite", "Homo sapiens" }
set IQ dimensions S = { "A": "accuracy", "R": "representation", "O": "objectivity", "C": "completeness" }
set negative levels of correctnes L- = [ 1, 3 ]
set positive levels of correctnes L+ = [ 5, 7 ]
set stories I = [ 1, 4 ]
select 1 random topic t ∈ T
for each IQ dimension S // Retrieve 24 statements (6 for each S) for a given topic t
  retrieve statements S1(t), S2(t) and S3(t) with l ∈ L-
  retrieve statements S5(t), S6(t) and S7(t) with l ∈ L+
for each IQ dimension S // Select 16 random statements (4 for each S) for story construction
  select 2 random statements with l ∈ L-
  select 2 random statements with l ∈ L+
for each story i ∈ I // Construct 4 random stories consisting of 1 statement for each S ∈ D
  for each IQ dimension S
    select 1 random statement
for each story i ∈ I // Participant evaluates the correctnes of each story i ∈ I and IQ dimension S ∈ D
  for each IQ dimension S
    user evaluates information quality IQ_init(t, S, i) 
    of a statement S for a given topic t as part of a story i

Every participant was randomly assigned one selected topic \(t\) from the pool of three available topics (one familiar and two unfamiliars to participants): the Occitan language, Alexandrite, and Homo sapiens (see step 6 in Algorithm 1 and activity A1 in Figure 4.1), as a source for the stories. Homo sapiens topic was one that participants were familiar with. Two other topics, with which the majority of participants were unfamiliar, were also selected. The lesser-known topic was either the mineral Alexandrite, an extremely rare gemstone, or the Occitan language, which fewer than \(1.5\) million people speak. Such a topic was intended to serve as a use case of how participants react when they come in contact with unknown, unverified information. Four different short stories were created from the selected topics for each participant, resulting in stories with the same topic, but IQ differing in each of the four dimensions.

For every selected topic \(t\) we prepared \(24\) statements with differing level of correctness (positive and negative) in terms of four IQ dimensions \(S\): objectivity, representation, accuracy and completeness (see steps 7–10 in Algorithm 1 and Table 4.2). The source for our statement construction was Wikipedia as it is open for everyone and easily accessible. It is also used as a general source of information in the majority of information quality-related research (Fidler and Lavbič 2017). For each of the four IQ dimensions \(S\), statements \(S_i\) were retrieved from the body text of the selected topic \(t\) (see activity A2 in Figure 4.1). The dimensions are represented by \(S \in \{ A, R, O, C \}\), with \(A\) (accuracy), \(R\) (representation), \(O\) (objectivity), and \(C\) (completeness), while the statements of dimensions are denoted by \(S_i\). Then, the level of correctness for each statement \(S_i\) was adjusted. The procedure for adjusting the level of correctness is described in Table 4.2. It is important to emphasize that each sentence focuses only on the main aspect of the selected IQ dimension (accuracy with interval of values, completeness with the number of included details, objectivity with the credibility of actor conveying information and representation with format and style consistency of minimal and maximal numbers, including units). The dimensions are represented by \(S \in \{ A, R, O, C \}\), with \(A\) (accuracy), \(R\) (representation), \(O\) (objectivity), and \(C\) (completeness), while the statements of dimensions are denoted by \(S_i\). All selected statements chosen for individual dimension \(S\) describe the same concept. However, each of the statements \(S_i\) have a different level of correctness. More precisely, the first three statements \(\Big(\{ S_1, S_2, S_3 \}\Big)\) have negative levels of correctness \(l^{-}\), while the latter three \(\Big(\{ S_5, S_6, S_7 \}\Big)\) have positive levels of correctness \(l^{+}\) (see steps 8–9 in Algorithm 1 and activity A3 in Figure 4.1). In Figure 4.1, statements with a negative level of correctness \(l^{-}\) are colored red, while statements with a positive level of correctness \(l^{+}\) are colored green. The lower the index of the statement, the more negative level of correctness it is associated with – that is, statement \(S_1\) is the least correct, while statement \(S_7\) is the most correct. The statement with the highest level of correctness \(l^{+} = 7\) was taken directly from the source. In the case of the topic Homo sapiens and IQ dimension accuracy \(A\), this statement \(A_7\) was “The height of an average human measures from 150 cm to 180 cm”. The following statement was then adjusted according to the rules presented in Table 4.2. Statement \(A_6\) \(\big(l^{+} = 6\big)\) decreased the range of accuracy by \(50\%\) and was defined as “The height of an average human measures from 157 cm to 173 cm”. Statement \(A_5\) \(\big(l^{+} = 5\big)\) increased the upper limit to a sum of maximum value and \(50\%\) range and was defined as “The height of an average human is less than 195 cm”. Statement \(A_3\) \(\big(l^{-} = 3\big)\) included only the upper limit and was defined as “The height of an average human is less than 180 cm”. Statement \(A_2\) \((l^{-} = 2)\) decreased the upper limit to \(50\%\) of the maximum value and was defined as “The height of an average human is less than 90 cm”. Statement with the lowest level of correctness \(A_1\) \(\big(l^{-} = 1\big)\) increased the upper limit to \(200\%\) of the maximum value and was defined as “The height of an average human is more than 360 cm”. This procedure of statement construction was performed for every IQ dimension \(S\) and topic \(t\).

Table 4.2: Details of statement construction. Statements with the level of correctness \(7\) are taken directly from the source, while statements with the level of correctness \(1\) are completely false.
Level of correctness Accuracy Completeness Objectivity Representation
7 true range value [min, max] true with all details most credible actor most consistent
6 range is decreased by 50% least important detail removed likely credible actor added ‘to’ and space ’ ’ before measurement unit
5 < max + 50% range most important detail removed slightly credible actor different number format
3 < max most important detail only slightly unlikely credible actor different number format and unformal measurement unit
2 < 50% max 2 least important details only unlikely credible actor different number format and wrong measurement unit
1 2 x max least important detail only not credible actor different number format, wrong numbers and measurement unit

Once we created the statements with the full set of levels of correctness, we selected actual statements for the assessment process. In Figure 4.1, this step is denoted by activities A4, A5, and A6. For each IQ dimension \(S\), we selected four random statements, two with a positive level of correctness (see step 12 in Algorithm 1) and two with a negative level of correctness (see step 13 in Algorithm 1). We selected statements from the set with positive \(l^{+}\) and set with negative statements \(l^{-}\) that were prepared earlier in the process. For example, in Figure 4.1 there are two randomly selected statements \(S_1\) and \(S_2\) with a negative level of correctness (see activity A6) and two random statements \(S_5\) and \(S_7\) with a positive level of correctness (see activity A5). The selection process was then performed for the remaining IQ dimensions \(S \in \{ A, R, O, C \}\) (see step 11 in Algorithm 1). Up to this point in the process, we selected \(16\) random statements, four per IQ dimension, half with negative and half with positive levels of correctness, all on the same topic \(t\). These statements served as the source for creating the four short stories that participants assessed.

Steps in conducting an experiment

Figure 4.1: Steps in conducting an experiment

The last step before participants start the assessment process involved story construction. From the statements, prepared in the previous step, we created four short paragraph-size stories. In Figure 4.1, this step is denoted by activities A7 and A8. Each story consists of four randomly chosen statements, where each statement belongs to one IQ dimension (see example in Table 4.3). We created stories progressively. For the \(i\)-th story (see step 15 in Algorithm 1), we first randomly selected an initial IQ dimension \(S\). If the \(i\)-th story did not include a statement for selected dimension \(S\) (see step 16 in Algorithm 1), we selected a random statement from among that had not been included in some previous story (see step 17 in Algorithm 1). Otherwise, we selected the next dimension from the set \(\{ A, R, O, C \}\). The story was complete when it contained exactly one statement from each IQ dimension – a total of four statements. The level of correctness of particular statements in the constructed story was unknown to the assessor. When all the prepared statements were included in the stories, they were ready for the assessment.

Table 4.3: An example of a generated story (see first story in Figure 4.1).
ID Dimension Level of correctness Content of statement Statement construction details
A2 Accuracy 2 The height of an average human is less than 90 cm. Less than 50% of maximum value, where average human measures from 150 cm to 180 cm.
R5 Representation 5 An average human weighs from 55 kg - 83 kilograms. Different number format, where most consistent form is 55 kg - 83 kg.
O6 Objectivity 6 Environmentalists are warning that expansion of human population is affecting the environment and survival of other plant and animal species. Environmentalists are likely credible actor in comparison to most credible actor - scientists.
C2 Completeness 2 In general, the human body consists of arms and legs. Two least important details only, where correct statement includes head, torso, pair of arms and legs.

The final step of the experiment was the evaluation phase (see activity A9 in Figure 4.1), in which participants assessed the stories. Specifically, each participant rated four short stories (see step 20 in Algorithm 1), each one paragraph long. Each paragraph is devoted to one topic and covers all four quality dimensions. Participants evaluated the correctness of statements in each paragraph (related to various IQ levels, see step 21 in Algorithm 1) using a \(7\) point Likert scale (from strongly disagree (1) to strongly agree (7)). For every story, participants had to decide how strongly they agree with perceived IQ for selected dimensions as defined in Table 4.1.

5 Findings

5.1 Results

As depicted in Figure 5.1, participants in our study successfully identified the differences between fake and real statement types throughout all considered IQ dimensions, although variations are significantly different across IQ dimensions.

Successful identification of fake and real statement types

Figure 5.1: Successful identification of fake and real statement types

To confirm that perceived IQ assessment \(\boldsymbol{IQ_d^t}\) is dependent on a statement type \(\boldsymbol{t} \in \{\text{fake}, \text{real}\}\) for every considered IQ dimension \(\boldsymbol{d} = \{ A, R, O, C \}\), we performed unpaired two-sample Wilcoxon or Mann–Whitney tests to compare two independent groups of samples. Our alternative hypothesis was \(\overline{IQ_d^{fake}} < \overline{IQ_d^{real}}\), with \(\alpha = 0.05\).

Participants were most successful in identifying the differences between fake and real information with the IQ dimension of objectivity, where the mean difference in perceived IQ assessments was calculated as \(\Delta_O^{\text{real} - \text{fake}} =\) \(\overline{IQ_{\text{O}}^{\text{real}}} - \overline{IQ_{\text{O}}^{\text{fake}}} =\) \(4.82 - 3.37 = \boldsymbol{1.45}\); the results are statistically significant \(\Big(W = 3.14\times 10^{5}\), \(p = 3.52\times 10^{-95}\Big)\). The effect of the remaining IQ dimensions (all with statistically significant differences) were as follows: representation with \(\Delta_R^{\text{real} - \text{fake}} = \boldsymbol{1.28}\) \(\Big(W = 3.5\times 10^{5}\), \(p = 1.24\times 10^{-75}\Big)\); completeness, \(\Delta_C^{\text{real} - \text{fake}} = \boldsymbol{1.05}\) \(\Big(W = 3.87\times 10^{5}\), \(p = 1.51\times 10^{-57}\Big)\); and accuracy, \(\Delta_A^{\text{real} - \text{fake}} = \boldsymbol{0.85}\) \(\Big(W = 4.18\times 10^{5}\), \(p = 8.68\times 10^{-45}\Big)\).

Figures 5.2 and 5.3 depict the influence of the independent variables of education, domain knowledge, and Big Five personality factors. The values under investigation and depicted in the figures are \(\Delta^{\text{real} - \text{fake}}\). The impact of individual values regarding the mean value for a given IQ dimension is also marked with a coloured background.

Influence of education and domain knowledge on identification of fake and real statement types. Domain knowledge includes two groups: Homo sapiens (well-known topic) and Alexandrite & the Occitan language (less-known topic). Education is grouped into below V (levels 0–3, from early childhood, primary, to lower and and upper secondary education), V (level 4, post-secondary non-tertiary education) and above V (levels 5–8, from short-cycle tertiary, Bachelor’s, Master’s to Doctoral education), according to the ISCED 2011 classification.

Figure 5.2: Influence of education and domain knowledge on identification of fake and real statement types. Domain knowledge includes two groups: Homo sapiens (well-known topic) and Alexandrite & the Occitan language (less-known topic). Education is grouped into below V (levels 0–3, from early childhood, primary, to lower and and upper secondary education), V (level 4, post-secondary non-tertiary education) and above V (levels 5–8, from short-cycle tertiary, Bachelor’s, Master’s to Doctoral education), according to the ISCED 2011 classification.

Influence of the Big Five personality factors on the identification of fake and real statement types

Figure 5.3: Influence of the Big Five personality factors on the identification of fake and real statement types

Asterisks (*) in Figures 5.2 and 5.3 depicts differences \(\Delta^{\text{real - fake}}_{d}\Big(\text{group}\Big)\) of \(\text{real}\) and \(\text{fake}\) statement types for a given IQ dimension \(d\), and \(\text{group}\), greater than the mean value \(\Delta^{real-fake}_{d}\) of the difference of \(\text{real}\) and \(\text{fake}\) statement types and a given IQ dimension \(d\).

5.2 Discussion

The purpose of this study was to empirically test the extent to which users (i.e. information consumers) can identify fake news based on information distortion. We built our study on the assumption that fake news affects individual dimensions of IQ. To investigate this issue, we conducted a study in which participants assessed IQ in terms of selected dimensions of a source document in the form of short paragraphs. The dataset was carefully created according to the guidelines laid out by Fidler and Lavbič (2017).

Regarding the first research question (RQ1), asking if users can perceive the IQ of short news and identify fake news, our results show that users are quite successful at identifying fake news. On average, they succeeded in detecting fake news on all IQ dimensions under study. The results depicted in Figure 5.1 demonstrate that the difference between false and real statement types is greater than zero \(\big(\Delta^{\text{real - fake}}_d > 0\big)\) for all dimensions \(\big(\forall d \in \{A, R, O, C\}\big)\), where \(\Delta^{\text{real - fake}}_d = \overline{IQ^{real}_d} - \overline{IQ^{fake}_d}\).

The second research question (RQ2) focused on the detection capability of individual IQ dimensions. Our results show that there are major differences between IQ dimensions. While users are able to detect fake news more easily using some IQ dimensions, they may struggle with other dimensions. Participants were most successful in detecting differences in the IQ dimension objectivity \(\big(\Delta^{\text{real - fake}}_R = 1.45\big)\), followed by representation \(\big(\Delta^{\text{real - fake}}_R = 1.28\big)\), completeness \(\big(\Delta^{\text{real - fake}}_R = 1.05\big)\), and accuracy \(\big(\Delta^{\text{real - fake}}_R = 0.85\big)\).

Finally, the third (RQ3) and fourth (RQ4) research questions investigated the factors that influence fake news detection. We found that users’ performance in fake news detection is influenced by several factors, such as knowledge of the problem domain, education, and users’ personality traits.

RQ3 deals with the impact of domain knowledge and education on the ability to more efficiently identify fake news. As depicted in Figure 5.2, users with knowledge of the problem domain (Homo sapiens) were more successful in detecting fake news than users with deficient domain knowledge (Alexandrite and the Occitan language). A similar conclusion about the impact of knowing the problem domain was reached by Xu et al. (2019). For the dimensions of representation, completeness, and accuracy \(\big(d_{RQ3} \in \{R, C, A\}\big)\), participants of our study achieved higher than average score \(\Delta_{d_{RQ3}}^{\text{real - fake}}\). For the dimension objectivity \(\big(d_{RQ3} = O\big)\), users achieved higher scores identifying fake news than average \(\Delta_O^{\text{real - fake}}\) if their level of education was V or greater. Only users with education lower than V struggled to successfully identify fake news, with scores \(\Delta_O^{\text{real - fake}}\) below average. In this case, foreknowledge itself is not enough, and more comprehensive knowledge is required to be successful in identifying fake news. We also observe that among users with foreknowledge of the problem domain, results for all four IQ dimensions \(d \in \{A, R, O, C\}\) consistently increase with a higher degree of education, meaning that education positively correlates with the ability to identify fake news. That is in line with the findings of Deme and Mahmoud (2020), who found similar effects concerning the impact of education on a different problem domain.

The fourth research question (RQ4) focused on which personal traits facilitate the detection of fake news and which IQ dimension is affected the most by these traits. As depicted in Figure 5.3, we used the IQ instrument to measure the influence of Big Five personality traits on the ability to detect fake news across all IQ dimensions. For each personality trait, we examined how the positive and negative features of this human trait affect the ability to detect the IQ of content and thus fake news. If the result \(\Delta_{d_{RQ4}}^{\text{real - fake}} \big(f, g\big)\) for a given IQ dimension \(d_{RQ4} \in \{A, R, O, C\}\), Big Five personality factor \(f \in \{\) Extraversion, Neuroticism, Conscientiousness, Agreeableness, Openness to Experience \(\}\) and Big Five personality group \(g \in \{\) negative, neutral, positive \(\}\) was greater than the average value \(\Delta_{d_{RQ_4}}^{\text{real - fake}}\) for the IQ dimension \(d_{RQ4}\), then users with particular personal trait \(f\) and particular attitude of this trait \(g\) were more successful in detecting the difference among the IQ of real and fake statements. The background colour (from red to green) of \(\Delta_{d_{RQ4}}^{\text{real - fake}} \big(f, g\big)\) values in Figure 5.3 depicts the magnitude of a difference of the mean value \(\Delta_{d_{RQ4}}^{\text{real - fake}} \big(f, g\big) - \Delta_{d_{RQ_4}}^{\text{real - fake}}\).

The results in Figure 5.3 confirm that IQ dimension accuracy caused the most problems in identifying fake news, regardless of personality traits. Objectivity was the IQ dimension by which users most easily distinguished real from fake news, consistent with RQ2 and Figure 5.1.

As depicted in Figure 5.3, individuals with pronounced positive conscientiousness \(\big(f^{\text{C}} =\) Conscientiousness and \(g^{+} =\) positive \(\big)\) were the most successful in identifying the differences between fake and real information \(\Delta^{\text{real - fake}}\), as they were able to identify fake news with all four IQ dimensions – objectivity \(\big(\Delta_{\text{O}}^{\text{real - fake}} \big(f^{\text{C}}, g^{+}\big) = 1.69\big)\), representation \(\big(\Delta_{\text{R}}^{\text{real - fake}} \big(f^{\text{C}}, g^{+}\big) = 1.44\big)\), completeness \(\big(\Delta_{\text{C}}^{\text{real - fake}} \big(f^{\text{C}}, g^{+}\big) = 1.22\big)\) and accuracy \(\big(\Delta_{\text{A}}^{\text{real - fake}} \big(f^{\text{C}}, g^{+}\big) = 0.96\big)\). These results are higher than average values \(\Delta_{d_{RQ4}}^{\text{real - fake}}\) for each dimension \(d_{RQ4}\). Although users had some difficulties identifying fake news by IQ dimension accuracy, their result \(\Delta_{\text{A}}^{\text{real - fake}}\big(f^{\text{C}}, g^{+}\big)\) was still the highest among all groups \(\big(f, g\big)\) and above the average \(\Delta_{\text{A}}^{\text{real - fake}}\). According to the Big Five personality model (Gosling, Rentfrow, and Swann Jr 2003; Woods and Hampson 2005), conscientiousness is the personality trait of being honest, organized, and hardworking. Sun, Kaufman, and Smillie (2018); Maier et al. (2020) showed that among the five major factors, conscientiousness contributes the most to success in school and at work. Conscientious people usually achieve a higher level of education (Schneider and Preckel 2017), which contributes to possessing more extensive knowledge; thus, they are more able to successfully detect misinformation. Our results confirm these findings from the literature that conscientiousness is positively correlated with education. The proportion of a group of participants with a pronounced positive conscientiousness (compared to neutral and negative groups) increased consistently from \(27\%\) for education group below V, to \(43\%\) for education group V and \(50\%\) for education group above V. This trend was also identified in domain knowledge, where the proportion of a group of participants with a pronounced positive conscientiousness increased from \(41\%\) for a low foreknowledge group (Alexandrite & the Occitan language) to \(48\%\) for a high foreknowledge group (Homo sapiens). This influence of education was also identified in fake news detection in our RQ3. For all four IQ dimensions, there is also a growing trend in the ability to identify fake news among participants with a pronounced negative \(\big(\Delta_{d_{RQ4}}^{\text{real - fake}}\big(f, g^{-}\big)\big)\) conscientiousness compared to those with a pronounced positive \(\big(\Delta_{d_{RQ4}}^{\text{real - fake}}\big(f, g^{+}\big)\big)\) value (see Figure 5.3). These findings confirm that conscientiousness facilitates the detection of fake news in terms of all four IQ dimensions.

The second most successful group of participants in fake news detection from results in Figure 5.3 were the negative extroverts or introverts \(\big(f^{\text{E}} =\) Extraversion and \(g^{-} =\) negative \(\big)\). They successfully detected the differences between fake and real information with IQ dimensions of objectivity \(\big(\Delta_{\text{O}}^{\text{real - fake}} \big(f^{\text{E}}, g^{-}\big) = 1.67\big)\), followed by representation \(\big(\Delta_{\text{R}}^{\text{real - fake}} \big(f^{\text{E}}, g^{-}\big) = 1.5\big)\), and completeness \(\big(\Delta_{\text{C}}^{\text{real - fake}} \big(f^{\text{E}}, g^{-}\big) = 1.19\big)\). Although Ehrenberg et al. (2008) and Correa, Hinsley, and De Zuniga (2010) found that extroverted people make more use of instant messaging services, our results show that the same group does not perform well at detecting fake news. Introverts prefer professions where they spend more time alone, e.g. as artists, mathematicians, engineers and researchers (Sun, Kaufman, and Smillie 2018; Maier et al. 2020). People who want to pursue such professions are usually enrolled in education for longer and gain more knowledge, much like conscientious individuals.

Individuals with pronounced negative neuroticism \(\big(f^{\text{N}} =\) Neuroticism and \(g^{-} =\) negative \(\big)\) were most successful at identifying fake news by the dimension objectivity \(\big(\Delta_{\text{O}}^{\text{real - fake}} \big(f^{\text{N}}, g^{-}\big) = 1.61\big)\). According to the Big Five personality model, these people are generally emotionally stable, confident, reliable, realistic, and practical (Sun, Kaufman, and Smillie 2018; Gosling, Rentfrow, and Swann Jr 2003). Interestingly, individuals with pronounced positive neuroticism \(\big(f^{\text{N}} =\) Neuroticism and \(g^{+} =\) positive \(\big)\) identified fake news quite successfully by the dimension completeness \(\big(\Delta_{\text{C}}^{\text{real - fake}} \big(f^{\text{N}}, g^{+}\big) = 1.21\big)\). According to the Big Five personality model, these are anxious people, complicating, emotional, sensitive, and not self-confident (Sun, Kaufman, and Smillie 2018; Gosling, Rentfrow, and Swann Jr 2003). Ehrenberg et al. (2008) found that individuals with high neuroticism make more use of instant messaging services. Furthermore, authors in Correa, Hinsley, and De Zuniga (2010) claim that in case of neuroticism, there is an association with instant messaging when compared to face-to-face interaction because instant messaging permitted additional time to contemplate responses. A similar conclusion was outlined by Ryan and Xeno (2011), where individuals, who are high in neuroticism, are more likely to prefer using the Wall on Facebook, which offers people with neurotic tendencies the opportunity to take their time formulating messages and responses. Similarly, Ma and Leung (2019) confirmed that neurotic people used LinkedIn for information purposes more, where they reacted more to professional information and also follow professional information more frequently. As individuals with pronounced positive neuroticism tend to use social media more, they process a higher share of content posted on social media than other users. Previous work has determined that they also take more time processing the content, which could explain their higher success rate in detecting false information by the dimension of completeness.

As depicted in Figure 5.3, participants with both, positive and negative attitudes of Agreeableness and Openness to Experience achieved above-average results for the IQ dimensions of Objectivity and Representation. Thus, we can conclude that these two personality traits do not have a meaningful influence on the detection of fake news in terms of IQ dimensions of Objectivity and Representation. In the case of IQ dimensions of Completeness and Accuracy, participants with pronounced positive Agreeableness scored only slightly above average, but the trend from negative to positive attitude for both IQ dimensions is increasing, which leads to the conclusion that Agreeableness may also facilitate the detection of fake news via the IQ dimensions of Completeness and Accuracy.

Based on our results and regarding RQ4, we can conclude that participants with pronounced positive conscientiousness are the most successful in detecting fake news. They consistently outperformed in the detection of fake news using all four IQ dimensions. The second most successful were introverts, outperforming the average detection of fake news in three IQ dimensions, and the third most successful were negative and positive neurotics, in the IQ dimensions of objectivity and completeness, respectively.

Our findings may have important implications for today’s world. Overall, research has found that the volume of fake news online is rising, and users are increasingly left to extract the truth themselves (Kim and Dennis 2018). In addition to intentional fabrication, sensationalism and competition for attention can also produce misinformation (Y. Wang et al. 2019). However, our findings are not in line with those of Moravec, Minas, and Dennis (2018). They found that social media users are poor at separating fake news from real news. Their work found that a fake news flag did not affect users’ judgements about the truth when news challenged their opinions. We speculate that such a scenario holds for convinced users who are not interested in the truth, and we believe that such a scenario does not apply to our work.

Our findings can also be connected to recent research that individuals who overclaim their level of knowledge also judge fake news to be more accurate (Pennycook and Rand 2020). The present study shows that education positively affects fake news detection. Previous work has shown that lack of knowledge is positively correlated with falling for fake news, and we showed that formal education improves the detection of false statements. The fact that domain knowledge has a huge impact on detection score is also in line with the findings of Pennycook and Rand (2020).

5.3 Practical implications

This study has several important implications for practice. First, our approach offers end-user an instrument for the manual assessment of shorter paragraphs for detection of misinformation. We believe that our approach can represent a valuable alternative to automatic approaches by exploiting different characteristics of false information. Automated approaches could have deficiencies such as a lack of high-quality labelled datasets, a lack of deep meaning, or inefficiencies in the early detection of misinformation (Shu et al. 2017; Zhou and Zafarani 2020).

Second, information consumers must critically evaluate the information quality of online sources. Our study has shown that information consumer possesses the knowledge required to identify misinformation but need to be aware that critical appraisal is his/her responsibility. Our findings are also in line with those of Y. Wang et al. (2019), who stated that instead of retracting misinformation, equipping individuals with the faculty for a critical assessment of the credibility of information might yield better results.

Third, our results show that there are user groups among the general population that are more effective than average at detecting misinformation. We identified the personality traits that allow more effective manual detection of misinformation. That suggests we should consider the findings when selecting the right candidate for a job that requires manual detection of fake news. Based on this finding, we could greatly complement and improve existing systems for automatic fake news detection.

Fourth, we suggest greater caution when consuming news from different sources for users who do not have pronounced outlined personality traits or do not have a higher level of education, or are not familiar with the problem domain.

Fifth, in crowd-sourced fact-checking, assessment relies on a large population of fact-checkers that resemble end-users rather than expert-based checkers (Zhou and Zafarani 2020). We believe that our approach could empower crowd-sourced fact-checkers for better misinformation detection in a cost-effective manner.

5.4 Limitations

Although the current study contributes to previous research, it has some limitations. First, our study investigated paragraph-size documents, like those typically found in microblogs. We selected paragraph-size documents because microblogs are ideal platforms for information dissemination within a short period (Yin, Zhang, and Liu 2020), and therefore, the diffusion of possible fake news. The effect of story size on IQ evaluation is addressed by Fidler and Lavbič (2017), where they employed Gricean principles to shorten Wikipedia articles to paragraph-size documents. They compared agreement level and overall perceived IQ by the IQ dimensions used in our research. Their results confirmed that shortening the unit of evaluation from full-size text to paragraph-size did not influence the agreement level of evaluations.

We should also be aware of the limitations that come from participants who assess individual quality dimensions – they may provide only a limited indication of the quality of the object (Arazy and Kopak 2011). This study focused on the detection of fake news by assessment of IQ. Such assessment is not an easy task, and Arazy, Kopak, and Hadar (2017) showed that achieving agreement among assessors can be challenging.

We are also aware that we could use additional problem domains next to those employed in our study. However, creating statements with a specific amount of misinformation by individual IQ dimensions requires lots of effort. As part of our study, participants self-assessed their knowledge of the narrower and broader areas of the given problem domains. The results obtained were consistent with our assumption that the domain Homo sapiens would be well known to them, while the other two were not.

Our study focused on fake news identification by considering various IQ dimensions. We employed a simplified method of acquiring information on participants’ personality traits, requesting users perform a simple self-assessment. This approach is in agreement with the use of the TIPI instrument, but we are aware that more robust approaches exist (e.g. BFI and NEO-PI-R). We did not utilize these comprehensive measurements of personality traits as they increase the payload to the participants who first have already agreed to invest substantial effort into a quality assessment of IQ dimensions.

6 Conclusion and future work

The responsibility for verifying content is increasingly being transferred from media sources to end-users on social media. We investigated how successful users are at detecting fake news in shorter content sources, namely paragraphs consisting of a few sentences. In terms of the factors that affect fake news detection, we examined the role of foreknowledge, education, and users’ personality traits. Our approach used IQ as a tool for source document assessment, which allowed us to identify how participants perceive various IQ dimensions. Personality traits were evaluated using a custom instrument based on the established TIPI instrument. We constructed an online questionnaire and collected data from a variety of users.

The results demonstrate that we can use IQ-based tools to successfully detect fake news. We found that participants successfully identified the difference between fake and real statements according to all IQ dimensions studied. We also found that domain knowledge affects fake news detection. As far as education is concerned, we showed that higher education in combination with domain knowledge increases the ability to detect fake news. Without domain knowledge, education does not always contribute to performance in detecting fake news.

Concerning personality traits, we found that conscientiousness contributes significantly to better fake news detection across all IQ dimensions.

In future work, we would like to replicate our study on an extended body of participants. For instance, it would be beneficial to focus on participants considered established information experts, such as librarians, journalists, and professors. Tweets and other short sources are only one source of fake news on the internet. To address this issue, we would also like to extend our study to even larger source documents.

Acknowledgments

We offer special thanks to Miloš Fidler, who was actively involved in the data collection process and contributed to the design in a manner that greatly assisted this research. We also thank all participants who generously shared their time for the purposes of data collection.

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

References

Agarwal, Nitin, and Yusuf Yiliyasi. 2010. “Information Quality Challenges in Social Media.” In The 15th International Conference on Information Quality, 234–48. Little Rock, Arkansas, USA, November 12-14, 2010: MIT Information Quality Program.
Ajao, Oluwaseun, Deepayan Bhowmik, and Shahrzad Zargari. 2018. “Fake News Identification on Twitter with Hybrid CNN and RNN Models.” In Proceedings of the 9th International Conference on Social Media and Society, 226–30. SMSociety ’18. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/3217804.3217917.
Allcott, Hunt, and Matthew Gentzkow. 2017. “Social Media and Fake News in the 2016 Election.” Journal of Economic Perspectives 31 (2): 211–36. https://doi.org/10.1257/jep.31.2.211.
Alrubaian, Majed, Muhammad Al-Qurishi, Mohammad Mehedi Hassan, and Atif Alamri. 2018. “A Credibility Analysis System for Assessing Information on Twitter.” IEEE Transactions on Dependable and Secure Computing 15 (4): 661–74. https://doi.org/10.1109/TDSC.2016.2602338.
Amichai-Hamburger, Yair, and Gideon Vinitzky. 2010. “Social Network Use and Personality.” Computers in Human Behavior 26 (6): 1289–95. https://doi.org/10.1016/j.chb.2010.03.018.
Arazy, Ofer, and Rick Kopak. 2011. “On the Measurability of Information Quality.” Journal of the American Society for Information Science and Technology 62 (1): 89–99. https://doi.org/10.1002/asi.21447.
Arazy, Ofer, Rick Kopak, and Irit Hadar. 2017. “Heuristic Principles and Differential Judgments in the Assessment of Information Quality.” Journal of the Association for Information Systems 18 (5): 1. https://doi.org/10.17705/1jais.00458.
Asudeh, Abolfazl, Hosagrahar Visvesvaraya Jagadish, You Wu, and Cong Yu. 2020. “On Detecting Cherry-Picked Trendlines.” Proceedings of the VLDB Endowment 13 (6): 939–52. https://doi.org/10.14778/3380750.3380762.
Bindu, PV, Rahul Mishra, and P Santhi Thilagam. 2018. “Discovering Spammer Communities in Twitter.” Journal of Intelligent Information Systems, 1–25. https://doi.org/10.1007/s10844-017-0494-z.
Bondielli, Alessandro, and Francesco Marcelloni. 2019. “A Survey on Fake News and Rumour Detection Techniques.” Information Sciences 497: 38–55. https://doi.org/10.1016/j.ins.2019.05.035.
C. Tandoc Jr., Edson, Zheng Wei Lim, and Richard Ling. 2018. “Defining Fake News.” Digital Journalism 6 (2): 137–53. https://doi.org/10.1080/21670811.2017.1360143.
Chang, Jiyoon, and Junseok Hwang. 2020. “The Role of Media in User Participation: Focusing on the Knowledge Activity in Online Space.” Telematics and Informatics, 101407. https://doi.org/10.1016/j.tele.2020.101407.
Chen, Yimin, Niall J Conroy, and Victoria L Rubin. 2015. “Misleading Online Content: Recognizing Clickbait as" False News".” In Proceedings of the 2015 ACM on Workshop on Multimodal Deception Detection, 15–19. https://doi.org/10.1145/2823465.2823467.
Conroy, Nadia K, Victoria L Rubin, and Yimin Chen. 2015. “Automatic Deception Detection: Methods for Finding Fake News.” Proceedings of the Association for Information Science and Technology 52 (1): 1–4. https://doi.org/10.1002/pra2.2015.145052010082.
Correa, Teresa, Amber Willard Hinsley, and Homero Gil De Zuniga. 2010. “Who Interacts on the Web?: The Intersection of Users’ Personality and Social Media Use.” Computers in Human Behavior 26 (2): 247–53. https://doi.org/10.1016/j.chb.2009.09.003.
Deme, Mamit, and Ali MA Mahmoud. 2020. “Effect of Quantity and Quality of Education on Per Capita Real-GDP Growth: Evidence from Low-and Middle-Income African Countries.” Applied Economics, 1–17. https://doi.org/10.1080/00036846.2020.1789058.
Ehrenberg, Alexandra, Suzanna Juckes, Katherine M White, and Shari P Walsh. 2008. “Personality and Self-Esteem as Predictors of Young People’s Technology Use.” Cyberpsychology & Behavior 11 (6): 739–41. https://doi.org/10.1089/cpb.2008.0030.
Ehrhart, Mark G, Karen Holcombe Ehrhart, Scott C Roesch, Beth G Chung-Herrera, Kristy Nadler, and Kelsey Bradshaw. 2009. “Testing the Latent Factor Structure and Construct Validity of the Ten-Item Personality Inventory.” Personality and Individual Differences 47 (8): 900–905. https://doi.org/10.1016/j.paid.2009.07.012.
Fallis, D. 2008. “Toward an Epistemology of Wikipedia.” Journal of the American Society for Information Science and Technology 59 (10): 1662–74. https://doi.org/10.1002/asi.20870.
Fidler, Miloš, and Dejan Lavbič. 2017. “Improving Information Quality of Wikipedia Articles with Cooperative Principle.” Online Information Review 41 (6): 797–811. https://doi.org/10.1108/OIR-01-2016-0003.
Gosling, Samuel D, Peter J Rentfrow, and William B Swann Jr. 2003. “A Very Brief Measure of the Big-Five Personality Domains.” Journal of Research in Personality 37 (6): 504–28. https://doi.org/10.1016/S0092-6566(03)00046-1.
Gravanis, Georgios, Athena Vakali, Konstantinos Diamantaras, and Panagiotis Karadais. 2019. “Behind the Cues: A Benchmarking Study for Fake News Detection.” Expert Systems with Applications 128: 201–13. https://doi.org/10.1016/j.eswa.2019.03.036.
Gustavsson, Mattias, and Carl Wänström. 2009. “Assessing Information Quality in Manufacturing Planning and Control Processes.” International Journal of Quality & Reliability Management. https://doi.org/10.1108/02656710910950333.
Hajli, Nick. 2015. “Social Commerce Constructs and Consumer’s Intention to Buy.” International Journal of Information Management 35 (2): 183–91. https://doi.org/10.1016/j.ijinfomgt.2014.12.005.
Iio, J. 2012. “Evaluating the Usefulness of Online Reviews.” In 2012 15th International Conference on Network-Based Information Systems, 855–60. https://doi.org/10.1109/NBiS.2012.30.
Ishida, Yoshiteru, and Sanae Kuraya. 2018. “Fake News and Its Credibility Evaluation by Dynamic Relational Networks: A Bottom up Approach.” Procedia Computer Science 126: 2228–37. https://doi.org/10.1016/j.procs.2018.07.226.
Jang, Wonseok Eric, Joon Sung Lee, Dae Hee Kwak, and Yong Jae Ko. 2019. “Meaningful Vs. Hedonic Consumption: The Effects of Elevation on Online Sharing and Information Searching Behaviors.” Telematics and Informatics 45: 101298. https://doi.org/10.1016/j.tele.2019.101298.
Kim, Antino, and Alan Dennis. 2018. “Says Who?: How News Presentation Format Influences Perceived Believability and the Engagement Level of Social Media Users.” In Proceedings of the 51st Hawaii International Conference on System Sciences. https://doi.org/10.24251/HICSS.2018.497.
Kshetri, Nir, and Jeffrey Voas. 2017. “The Economics of ‘Fake News’.” IT Professional 19 (6): 8–12. https://doi.org/10.1109/MITP.2017.4241459.
Kucharski, Adam. 2016. “Study Epidemiology of Fake News.” Nature 540 (7634): 525–25. https://doi.org/10.1038/540525a.
Lahuerta-Otero, Eva, and Rebeca Cordero-Gutiérrez. 2016. “Looking for the Perfect Tweet. The Use of Data Mining Techniques to Find Influencers on Twitter.” Computers in Human Behavior 64: 575–83. https://doi.org/10.1016/j.chb.2016.07.035.
Lamers, Sanne MA, Gerben J Westerhof, Viktória Kovács, and Ernst T Bohlmeijer. 2012. “Differential Relationships in the Association of the Big Five Personality Traits with Positive Mental Health and Psychopathology.” Journal of Research in Personality 46 (5): 517–24. https://doi.org/10.1016/j.jrp.2012.05.012.
Lazer, David MJ, Matthew A Baum, Yochai Benkler, Adam J Berinsky, Kelly M Greenhill, Filippo Menczer, Miriam J Metzger, et al. 2018. “The Science of Fake News.” Science 359 (6380): 1094–96. https://doi.org/10.1126/science.aao2998.
Lee, Y. W., D. M. Strong, B. K. Kahn, and R. Y. Wang. 2002. AIMQ: A Methodology for Information Quality Assessment.” Information & Management 40 (2): 133–46. https://doi.org/10.1016/S0378-7206(02)00043-5.
Lim, S. 2009. “How and Why Do College Students Use Wikipedia?” Journal of the American Society for Information Science and Technology 60 (11): 2189–2202. https://doi.org/10.1002/asi.21142.
Luyt, Brendan, and Daniel Tan. 2010. “Improving Wikipedia’s Credibility: References and Citations in a Sample of History Articles.” Journal of the American Society for Information Science and Technology 61 (4): 715–22. https://doi.org/10.1002/asi.21304.
Ma, Shyron Qianyun, and Louis Leung. 2019. “The Impacts of Personality Traits, Use Intensity and Features Use of LinkedIn on Bridging Social Capital.” Computers in Human Behavior 14 (4): 1059–78. https://doi.org/10.1007/s11482-018-9635-y.
Mai, Jens-Erik. 2013. “The Quality and Qualities of Information.” Journal of the American Society for Information Science and Technology 64 (4): 675–88. https://doi.org/10.1002/asi.22783.
Maier, Christian, Jens Mattke, Katharina Pflügner, and Tim Weitzel. 2020. “Smartphone Use While Driving: A Fuzzy-Set Qualitative Comparative Analysis of Personality Profiles Influencing Frequent High-Risk Smartphone Use While Driving in Germany.” International Journal of Information Management 55: 102207. https://doi.org/10.1016/j.ijinfomgt.2020.102207.
Malhotra, Bhaye, and Dinesh Kumar Vishwakarma. 2020. “Classification of Propagation Path and Tweets for Rumor Detection Using Graphical Convolutional Networks and Transformer Based Encodings.” In 2020 IEEE Sixth International Conference on Multimedia Big Data (BigMM), 183–90. IEEE. https://doi.org/10.1109/BigMM50055.2020.00034.
Meel, Priyanka, and Dinesh Kumar Vishwakarma. 2020. “Fake News, Rumor, Information Pollution in Social Media and Web: A Contemporary Survey of State-of-the-Arts, Challenges and Opportunities.” Expert Systems with Applications 153: 112986. https://doi.org/10.1016/j.eswa.2019.112986.
Mehdizadeh, Soraya. 2010. “Self-Presentation 2.0: Narcissism and Self-Esteem on Facebook.” Cyberpsychology, Behavior, and Social Networking 13 (4): 357–64. https://doi.org/10.1089/cyber.2009.0257.
Moravec, Patricia, Randall Minas, and Alan R Dennis. 2018. “Fake News on Social Media: People Believe What They Want to Believe When It Makes No Sense at All.” Kelley School of Business Research Paper, no. 18-87. https://doi.org/10.2139/ssrn.3269541.
Nakayama, Makoto, and Yun Wan. 2017. “Exploratory Study on Anchoring: Fake Vote Counts in Consumer Reviews Affect Judgments of Information Quality.” Journal of Theoretical and Applied Electronic Commerce Research 12 (1): 1–20. https://doi.org/10.4067/S0718-18762017000100002.
Newman, Nic, William Dutton, and Grant Blank. 2013. “Social Media in the Changing Ecology of News: The Fourth and Fifth Estates in Britain.” International Journal of Internet Science 7 (1).
Ozbay, Feyza Altunbey, and Bilal Alatas. 2020. “Fake News Detection Within Online Social Media Using Supervised Artificial Intelligence Algorithms.” Physica A: Statistical Mechanics and Its Applications 540: 123174. https://doi.org/10.1016/j.physa.2019.123174.
Paggi, Horacio, Javier Soriano, Juan A. Lara, and Ernesto Damiani. 2021. “Towards the Definition of an Information Quality Metric for Information Fusion Models.” Computers & Electrical Engineering 89: 106907. https://doi.org/10.1016/j.compeleceng.2020.106907.
Pennycook, Gordon, and David G. Rand. 2020. “Who Falls for Fake News? The Roles of Bullshit Receptivity, Overclaiming, Familiarity, and Analytic Thinking.” Journal of Personality 88 (2): 185–200. https://doi.org/10.1111/jopy.12476.
Plotkina, Daria, Andreas Munzel, and Jessie Pallud. 2018. “Illusions of Truth - Experimental Insights into Human and Algorithmic Detections of Fake Online Reviews.” Journal of Business Research. https://doi.org/10.1016/j.jbusres.2018.12.009.
Polasik, Michal, Agnieszka Huterska, and Andrzej Meler. 2018. “The Impact of Formal Education on Financial Inclusion in Payment Services.” E-MENTOR, no. 1: 30–40. https://doi.org/10.15219/em73.1336.
Qin, Yumeng, Wurzer Dominik, and Cunchen Tang. 2018. “Predicting Future Rumours.” Chinese Journal of Electronics 27 (3): 514–20. https://doi.org/10.1049/cje.2018.03.008.
Ruchansky, Natali, Sungyong Seo, and Yan Liu. 2017. “Csi: A Hybrid Deep Model for Fake News Detection.” In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, 797–806. CIKM ’17. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/3132847.3132877.
Ryan, Tracii, and Sophias Xeno. 2011. “Who Uses Facebook? An Investigation into the Relationship Between the Big Five, Shyness, Narcissism, Loneliness, and Facebook Usage.” Computers in Human Behavior 27 (5): 1658–64. https://doi.org/10.1016/j.chb.2011.02.004.
Sahoo, Somya Ranjan, and BB Gupta. 2021. “Multiple Features Based Approach for Automatic Fake News Detection on Social Networks Using Deep Learning.” Applied Soft Computing 100: 106983. https://doi.org/10.1016/j.asoc.2020.106983.
Schneider, Michael, and Franzis Preckel. 2017. “Variables Associated With Achievement in Higher Education: A Systematic Review of Meta-Analyses.” Psychological Bulletin 143 (6): 565–600. https://doi.org/10.1037/bul0000098.
Shu, Kai, Amy Sliva, Suhang Wang, Jiliang Tang, and Huan Liu. 2017. “Fake News Detection on Social Media: A Data Mining Perspective.” SIGKDD Explor. Newsl. 19 (1): 22–36. https://doi.org/10.1145/3137597.3137600.
Solnet, David J, Anna L Kralj, Elisa Moncarz, and Christine Kay. 2010. “Formal Education Effectiveness and Relevance: Lodging Manager Perceptions.” Journal of Hospitality & Tourism Education 22 (4): 15–24. https://doi.org/10.1080/10963758.2010.10696988.
Sun, Jessie, Scott Barry Kaufman, and Luke D Smillie. 2018. “Unique Associations Between Big Five Personality Aspects and Multiple Dimensions of Well-Being.” Journal of Personality 86 (2): 158–72. https://doi.org/10.1111/jopy.12301.
Viviani, Marco, and Gabriella Pasi. 2017. “Credibility in Social Media: Opinions, News, and Health Information—a Survey.” Wiley Interdisciplinary Reviews-Data Mining and Knowledge Discovery 7 (5). https://doi.org/10.1002/widm.1209.
Volkova, Svitlana, Kyle Shaffer, Jin Yea Jang, and Nathan Hodas. 2017. “Separating Facts from Fiction: Linguistic Models to Classify Suspicious and Trusted News Posts on Twitter.” In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 647–53. Vancouver, Canada: Association for Computational Linguistics. https://doi.org/10.18653/v1/P17-2102.
Vosoughi, Soroush, Mostafa Neo Mohsenvand, and Deb Roy. 2017. “Rumor Gauge: Predicting the Veracity of Rumors on Twitter.” ACM Trans. Knowl. Discov. Data 11 (4). https://doi.org/10.1145/3070644.
Vosoughi, Soroush, Deb Roy, and Sinan Aral. 2018. “The Spread of True and False News Online.” Science 359 (6380): 1146–51. https://doi.org/10.1126/science.aap9559.
Wang, Richard Y., and Diane M. Strong. 1996. “Beyond Accuracy: What Data Quality Means to Data Consumers.” Journal of Management Information Systems 12 (4): 5–33. https://doi.org/10.1080/07421222.1996.11518099.
Wang, William Yang. 2017. Liar, Liar Pants on Fire: A New Benchmark Dataset for Fake News Detection.” In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 422–26. Vancouver, Canada: Association for Computational Linguistics. https://doi.org/10.18653/v1/P17-2067.
Wang, Yuxi, Martin McKee, Aleksandra Torbica, and David Stuckler. 2019. “Systematic Literature Review on the Spread of Health-Related Misinformation on Social Media.” Social Science & Medicine 240. https://doi.org/10.1016/j.socscimed.2019.112552.
Wardle, Claire. 2017. “Fake News. It’s Complicated.” First Draft News 16: 1–11. https://firstdraftnews.org/latest/fake-news-complicated/.
Wilson, Kathryn, Stephanie Fornasier, and Katherine M White. 2010. “Psychological Predictors of Young Adults’ Use of Social Networking Sites.” Cyberpsychology, Behavior, and Social Networking 13 (2): 173–77. https://doi.org/10.1089/cyber.2009.0094.
Woods, Stephen A, and Sarah E Hampson. 2005. “Measuring the Big Five with Single Items Using a Bipolar Response Scale.” European Journal of Personality: Published for the European Association of Personality Psychology 19 (5): 373–90. https://doi.org/10.1002/per.542.
Wu, Liang, Fred Morstatter, Kathleen M Carley, and Huan Liu. 2019. “Misinformation in Social Media: Definition, Manipulation, and Detection.” ACM SIGKDD Explorations Newsletter 21 (2): 80–90. https://doi.org/10.1145/3373464.3373475.
Xu, Kuai, Feng Wang, Haiyan Wang, and Bo Yang. 2019. “Detecting Fake News over Online Social Media via Domain Reputations and Content Understanding.” Tsinghua Science and Technology 25 (1): 20–27. https://doi.org/10.26599/TST.2018.9010139.
Yaari, E., S. Baruchson-Arbib, and J. Bar-Ilan. 2011. “Information Quality Assessment of Community-Generated Content - A User Study of Wikipedia.” Journal of Information Science 37 (5): 487–98. https://doi.org/10.1177/0165551511416065.
Yin, Chunxiao, Xiaofei Zhang, and Libo Liu. 2020. “Reposting Negative Information on Microblogs: Do Personality Traits Matter?” Information Processing & Management 57 (1): 102106. https://doi.org/10.1016/j.ipm.2019.102106.
Zhang, Hu, Zhuohua Fan, Jiaheng Zheng, and Quanming Liu. 2012. “An Improving Deception Detection Method in Computer-Mediated Communication.” Journal of Networks 7 (11): 1811. https://doi.org/10.4304/jnw.7.11.1811-1816.
Zhang, Xichen, and Ali A. Ghorbani. 2020. “An Overview of Online Fake News: Characterization, Detection, and Discussion.” Information Processing & Management 57 (2). https://doi.org/10.1016/j.ipm.2019.03.004.
Zhou, Xinyi, and Reza Zafarani. 2020. “A Survey of Fake News: Fundamental Theories, Detection Methods, and Opportunities.” ACM Computing Surveys (CSUR) 53 (5): 1–40. https://doi.org/10.1145/3395046.
Zhu, Linlin, He Li, Wu He, and Chuang Hong. 2020. “What Influences Online Reviews’ Perceived Information Quality?” The Electronic Library. https://doi.org/10.1108/EL-09-2019-0208.
Zhu, Linlin, He Li, Feng-Kwei Wang, Wu He, and Zejin Tian. 2020. “How Online Reviews Affect Purchase Intention: A New Model Based on the Stimulus-Organism-Response (SOR) Framework.” Aslib Journal of Information Management. https://doi.org/10.1108/AJIM-11-2019-0308.
Zubiaga, Arkaitz, Ahmet Aker, Kalina Bontcheva, Maria Liakata, and Rob Procter. 2018. “Detection and Resolution of Rumours in Social Media: A Survey.” ACM Computing Surveys (CSUR) 51 (2): 1–36. https://doi.org/10.1145/3161603.