Next Article in Journal
Urban Traffic Signal Control under Mixed Traffic Flows: Literature Review
Previous Article in Journal
Application of Machine Learning Algorithms for the Validation of a New CoAP-IoT Anomaly Detection Dataset
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Digital Information Credibility: Towards a Set of Guidelines for Quality Assessment of Grey Literature in Multivocal Literature Review

1
Department of Computer Software Engineering, University of Engineering and Technology, Peshawar 25120, Pakistan
2
Department of Computer Software Engineering, University of Engineering and Technology, Mardan 23200, Pakistan
3
Department of Computer Science, University of Engineering and Technology, Mardan 23200, Pakistan
4
EIAS Data Science and Blockchain Lab, College of Computer and Information Sciences, Prince Sultan University, Riyadh 11586, Saudi Arabia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(7), 4483; https://doi.org/10.3390/app13074483
Submission received: 2 March 2023 / Revised: 24 March 2023 / Accepted: 24 March 2023 / Published: 1 April 2023

Abstract

:
Credibility, in general, can be interpreted as a sense of trust in someone. The credibility of information remarkably influences the public’s willingness to do or not to do some things. In this research study, the credibility of digital news stories can be interpreted as the sense of confidence a person has in a source of available information that affects their behavior. Humans spread less credible information instead of more credible information very quickly because humans take an interest in fear, disgust, and surprise. Less credible news may affect individuals as well as economies. Therefore, there is a dire need in the current digital era to find out what affects the credibility of digital news stories. This study aims to review the published literature and the grey literature to determine the factors affecting digital news credibility and the factors that build credibility in digital news stories. In this paper, we have developed a multivocal literature review protocol to assess the credibility of digital news stories. The multivocal literature review is an advanced version of the systematic literature review that searches for grey literature in addition to the published literature. The expected outcomes after implementing our protocol will be a list of credibility factors and their practices that can play a vital role in ensuring the credibility of digital news stories. Based on this protocol, we formulated guidelines for the quality assessment of grey literature. The future direction of this research is to analyze the factors through multi-criteria decision-making (MCDM), i.e., analytical hierarchical process.

1. Introduction

The World Wide Web (WWW) is the medium of information used by institutions [1]. The information about these institutions is continuously growing. Regardless of the physical location, the World Wide Web quickly shares information. The increase in the consumption of this information rapidly changes with the addition of new Web pages, with a wide range of indexing each day [2]. The rapid changes in technologies and the short lifespan of digital objects compel the researcher to find ways out of this predicament. Hence, there is a dire need to preserve the information on the WWW for future generations. The preservation of information on the WWW is known as digital preservation [3,4,5,6,7]. The information to be stored for future generations may be of any nature, i.e., less-credible information or only information in the literature. We investigated several articles and found that these studies begin with the definition of the term credibility. According to the dictionary, “credible” is defined as “is worthy of confidence, plausible, or believable”. This word is derived from the Latin word “cedere”. Its meaning is to “believe” [8]. Therefore, it can be simply defined as truthfulness or believability [9]. It is closely related to authority, accuracy, quality, and reliability [10]. The general definition of the credibility of information is the believability of information [1]. The believability depends on the individual’s perception, which cannot always separate less credible information from more credible information. Therefore, a major question arises. What content of whom do you trust? [11]. Information in the form of news makes the situation more interesting. Less credible news affects individuals and economies [12,13]. These rumors have huge impacts on the rise and fall of stock prices and large-scale investments. The inaccurate information has disrupted our responses to everything from terrorist attacks to natural disasters [10,14]. We studied many multivocal literature reviews (MLRs) protocols and their guidelines [15,16,17] to develop the protocol presented in this paper.

Motivation and Novelty

News information credibility is as important as general information credibility and becomes more important if the news is generated from multiple sources instantly and continuously. It is necessary to examine the reliability of news articles based on different criteria. The criteria need to be devised using an enhanced systematic approach. One of the primary motivations of this research is to find out the factors that affect the credibility of digital news stories. Mainstream news reflects the traditions, ethics, and skills of society. Preserving credible news will help future generations understand the culture and society of our era. However, such preservation is difficult in the current digital era as we are constantly being bombarded with both credible and non-credible information. To determine which sources are credible and therefore need to be preserved, we need reliable metrics. This study aims to develop these metrics. The contributions of this study are as follows:
  • In this study, we propose a set of guidelines for the assessment of unpublished sources, which will help the researchers to adopt a multivocal literature review.
  • To evaluate the authenticity of digital news stories, we have devised a multivocal literature review protocol.

2. Literature Review

Educators, librarians, and application providers are interested in spreading information as much as possible. They are often not concerned with its nature and whether or not it is credible [18]. The nature of the information is essential. We are interested in promoting the credibility of information instead of spreading rumors. However, this does not receive much attention as it is the end user’s problem [19]. Finding credible information has become a challenging computational task in an atmosphere with so much non-credible information available. Researchers are now focusing on the nature of information rather than just the information itself. This different thought process is a huge step forward in research into credible information. Digital information media, such as social technologies, play a vital role in spreading less credible information. These technologies facilitate the information providers because they are for the most part not peer-reviewed [20].
News makes the situation more interesting. Whether political, social, sports, entertainment, or satirical news, the news can be less credible or more credible. The distribution of news generally provides its sources. Often, people do not understand the context of the stories they read, but they keep forwarding the information to others.
To fully overcome the lack of credibility in news stories, we first need to determine what is less credible and more credible. This can happen if we examine the diffusion in-depth after differentiating the credible information from the less credible information. This study aims to determine how to distinguish these factors and model best-practices for their implementation. The increase in the number of websites yields an overwhelming variety and volume of content. However, content credibility is a significant question in many situations. The importance of content is based on the increase or decrease of the credibility of the content. The credibility assessment can be incorrect without making the groups of contents, i.e., Article, Discussion Source, Help, etc. [21].
With the emergence of Web 2.0 and social media, the contents of the internet have become the primary source of spreading information. This information boom leads to the tsunami of Web pages having so many news stories, audio, and videos. This information is both less credible and more credible. Therefore, the trend of research from different aspects of content switched to the “credibility of the contents”. The words “credibility” and “Trust” are used interchangeably [12]. Many researchers called the word credibility also believe-ability. Thus, Believable information is credible information. People understand credibility by employing more than one dimension concurrently. All the reviewed research [22,23,24,25] highlights the need for powerful credibility assessment abilities among the users. For example, how do people assess information on the Web? Are they aware of potential bias and vested pursuit of content writers as they determine whether the information is appropriate, correct, or plausible? The problem is particularly vital as, with the exponential proliferation of information on any subject matter over the Web, users emerge as the arbiters of records accuracy in a domain wherein they are informed and in environments in which they may be no longer.
The digital information viewer is facing a significant challenge in the credibility of content enhancement. The Researchers from STUP- Stanford University Persuasive Technology Lab have published a series of publications related to Web content credibility [9,26]. In his article [14], Fog B investigates how different links and domains affect the perception of people’s in-terms of credibility with specific implications. Some of the research studies are based on the effect of the credibility of a business or E-Commerce websites [26]. According to [20,27,28], several studies were reported from the literature on the credibility of advertisement in the field of marketing. These research articles analyze the contents and a few attributes of promotion. One of our research’s major primary goals is the factors that ensure credibility. In the literature, we found that the perception regarding the credibility of the advertisement includes the reputation of the company and its experience. The credibility of the information is measured via certain factors, i.e., the message contents, source credibility, source bias, and source reputation [29]. Medical information has received widespread attention due to its credibility issue. Almost 50% of Internet users seek health information on the Web. Due to the critical nature of the information, many researchers emphasize the need for a credibility assessment of health-related information on the Web [19,30].
The credibility of news can be measured by detecting less-credible news stories. Therefore, Less-Credible news detection can be defined as the provision of specific news articles, i.e., editorial, news report, expose, etc., intentionally deceived [31]. According to Chen, 55% of readers do not read the full article. They only click on the link and go through it without reading it. These types of readings can be considered “click-bait”. This click-bait plays a vital role in spreading less-credible information [32]. The popularity of social media platforms such as Twitter and Facebook pushed data to consumers in a speedy way. The share feature in these platforms makes the information available for all [33]. The standard-issue leads the consumer to pay less attention. While attention to detail becomes, the contents are de-contextualized from their source, and facts are mixed with fiction or even personal bias.

3. Methodology

We have adopted a multivocal Literature Review for this study. A comprehensive literature review (MLR) involves evaluating all available literature, including formal academic publications like journals and conference papers, as well as practitioner literature such as blog posts and white papers. Its purpose is to identify, analyze, and interpret phenomena of interest. MLR studies are particularly useful in Computer Science and Engineering where new advancements are taking place and academic research is limited, as they can provide significant advantages. The work presented in this document is motivated by a set of research questions that form the basis of our study:
RQ 1: What factors are identified in the literature (as well as in the grey literature) to be considered for ensuring digital information credibility of digital news stories?
RQ 2: What are the practices, as identified in the literature (as well as in the grey literature), for implementing the factors to ensure digital information credibility of digital news stories?

3.1. Constructing Search Terms

Population: Digital News Stories
Intervention: Credibility Factors, Characteristics, Practices.
Outcomes of relevance: To determine the Factors and Solutions to ensure credibility in digital news stories.
Experimental Design: Multivocal Literature Review.

3.2. Search Strategy

3.2.1. Trial Search

The search string was utilized to conduct a trial search on Google Scholar and Springer Link.
(“Digital news stories ” OR “information credibility”) AND (“credibility factors” OR “credibility indicators” OR “believability factors”)
We tried the search string in the aforementioned digital libraries and retrieved research publications. After testing the search string, we finalized it. This string will be further used to validate and develop practical search terms.

3.2.2. Identifying Search Terms

We used the following search approach to construct the search terms.
  • Use the Research Questions to derive significant terms by identifying population, intervention, and outcome.
  • Find alternative spellings and synonyms for the significant terms.
  • Verify the keywords in any relevant sources.
  • If your database allows it, use the ‘OR’ operator to concatenate alternate spellings and synonyms and the ‘AND’ operator to concatenate significant terms.

3.2.3. Result for the Search String

RQ: What are the factors, as identified in the literature as well in grey literature, to be considered for ensuring digital information credibility of digital news stories?
We formulated the search string as per the syntax/limitations of each digital library. These search strings are given in Table 1.

3.3. Resources to Be Searched

3.3.1. Search through Relevant Digital Libraries (Published Sources)

  • IEEE XPLORE
  • SCIENCE DIRECT
  • ACM
  • GOOGLE SCHOLAR (SEARCH ENGINE)
  • SPRINGER LINK

3.3.2. Search for Grey Literature

To locate documents in the grey literature, we used non-specialized internet search engines. Grey literature is a broad category of publications that includes a variety of report types, such as technical, research, project, and annual reports, as well as working papers, government documents, white papers, evaluations, and videos, among others. These materials are typically not published by commercial publishers and are not subject to their control. As recommended by Garousi et al. [17], grey studies can be identified by exploring search strings on search engines. We have studied a number of MLR protocols while designing this protocol [15,17,34,35,36,37,38,39,40]. We found MLR guidelines proposed by Garousi et al. [17] in the literature. The guidelines suggested two different strategies for searching, i.e., Automatic Search and Manual Search. Various search engines can do automatic Searches like Google search engine (https://www.google.com, accessed on 1 March 2023), Thesis Global databases and ProQuest Dissertations. Manual Search methods consist of an informal pre-search. The detail is given below.
Google provides numerous studies, and the sources should be limited to a specific size that can provide the best records and be easily readable. Only the first few pages are relevant to the study, while the rest are mostly irrelevant. This relevance can be achieved by using a page rank algorithm [41]. We will stop proceeding with more results once the page does not yield relevant items. Following the MLR guidelines [17], we will combine the inclusion and exclusion criteria for grey literature with quality assessment criteria.

3.4. Search Documentation

The data will be stored in a document to extract the final results. The search results will be recorded in the document provided in Table 2.

3.5. Search Result Management

The references for each search string result will be stored in separate folders. Screenshots of the search query results will also be taken and stored in the aforementioned folder to organize the records better. Additionally, the format shown in Table 3 will be used for the primary selection. Specifically, we will use the SD number of the article for Science Direct, the SL number of the article for SpringerLink, and the IEEE number of the article for IEEEXplore.
The ’paper ID’ is the number assigned to a particular paper by the search query after applying the search string to the digital library. The process of removing duplicated occurrences involves eliminating studies from one of the digital libraries if the same paper is found in more than one library.

4. Publication Selection of Published Sources

The selection of papers from a large number of papers is a very challenging task. We have adopted inclusion and exclusion criteria to select the most relevant articles and exclude irrelevant studies. These criteria specify that the paper should be in English and that the paper’s title should match our research title. Furthermore, if the paper’s title only partially matches, we will review the abstract and include it if it is relevant; otherwise, we will exclude it.

4.1. Inclusion Criteria for Published Sources

When applied to digital libraries, the search string will yield many studies. The data will be extracted from the sources once checked against the inclusion criteria. The sources will only be included if they are deemed relevant to our research questions. We will only consider sources related to the credibility of digital information in digital news stories. The criteria are listed below:
Sources That Explain:
  • The factors/motivators/success factors for digital information credibility.
  • The practices/solutions for the factors in digital information credibility.
  • The context of digital news stories for credibility purposes.
  • The context of digital news archives for credibility purposes.

4.2. Exclusion Criteria for Published Sources

Sources that are irrelevant to our research questions will not be included. The decision to exclude them is based on the exclusion criteria, which are explained below.
  • Sources that are not relevant to our research questions.
  • Sources that do not describe digital information credibility.
  • Sources that do not describe credibility in digital news stories.

4.3. Selecting Primary Sources

We are utilizing the tollgate approach proposed by Afzal et al. [42] to select the publications. The first phase of our source selection process involves reviewing the title, keywords, and abstract of primary sources. The purpose of this phase is to identify sources that are most relevant to our research questions and exclude irrelevant sources. The selection is further refined by reviewing the full text of studies and checking against the inclusion and exclusion criteria mentioned above. The sources will be sent to a secondary reviewer for further assessment against the inclusion and exclusion criteria to eliminate any remaining uncertainty. Records will be kept for each source selection to explain why a primary source was included or excluded from the final evaluation. Therefore, this section will be helpful in explaining the rationale behind the selection or removal of primary sources.

5. Publication Selection of Unpublished Sources

In the first step, the search string will be applied to all three sources specified previously in Section 3.3.2, namely Google, ProQuest, and a manual search for appropriate sources. After gathering studies using automatic and manual searches, the titles and abstracts will be thoroughly evaluated using the inclusion and exclusion criteria outlined in the following sections to eliminate irrelevant studies. If we encounter any difficulties in adding or removing a particular report, we will consult with our research supervisors to devise the best possible solution. The inclusion and exclusion criteria will be applied to the entire text of the pre-selected studies in the third stage.

5.1. Inclusion Criteria for Grey Literature

  • Ph.D./Master’s thesis, key consulting firms, and digital news agencies who have accompanying support resources on their websites (e.g., case studies, blogposts).
  • The study is relevant to the search terms.
  • The study is written in the English language.

5.2. Exclusion Criteria for Grey Literature

  • Studies that are not explicitly focused on digital news stories.
  • Studies that do not address the trustworthiness/credibility of news stories.
  • Studies for which the entire text is unavailable.
  • Duplicate research items (the same studies can be found in other journals).
  • Duplicate research items (the same studies in order, with only the most relevant or significant study included).

6. Publication Quality Assessment

Publication Quality Assessment for Published Literature

Quality assessment is a crucial activity that enhances the strength of a study. To improve the quality of our work, we have designed a short questionnaire to assess the quality of candidate papers. The questionnaire will be distributed via email among senior researchers and experts. After filling out the questionnaire, a scoring plan will be applied as elaborated below. This scoring plan is presented in columns A, B, C, D, E, and Score in Table 4. The mechanism for assessing the literature was suggested by Khan et al. [43] and [44]. We have added a few more checklists to this and proposed the following mechanism for evaluating the quality of the literature:
Has the work been accepted for publication in a reputable journal or conference? This question will be ranked using the Computing Research Education (CORE) 2018 and Journal Citation Reports (JCR) 2018 lists of computer science conferences. Additionally, all seminars, workshops, and symposiums have been categorized as follows:
  • To assess the ranking of a paper, it is necessary to consider whether it has been published in a recognized and stable journal or conference. The ranking will be determined by referring to the 2018 Computing Research Education (CORE) and Journal Citation Reports (JCR) lists. The conferences, seminars, workshops, and symposiums have been categorized into four groups based on their ranking: A1, A, B, and C. A conference that is not ranked is given a score of 0. The JCR provides rankings based on Impact Factor (IF), with Q1 denoting the top 25%, Q2 the top 50% to top 25%, Q3 the top 75% to top 50%, and Q4 the bottom 25%. A journal not included in the JCR list is given a score of 0.
  • The primary focus of the paper is to investigate the factors that contribute to the credibility of digital news stories. The score awarded is 1 for ”Yes”, 0.5 for ”Partially”, and 0 for ”No.”
  • The paper evaluates an approach to deal with the credibility factors of digital news stories. If the paper presents a new approach, it is given a score of 1 for ”Yes” 0.5 for ”Partially” if it offers an existing approach, and 0 for ”No” if it does not present a solution.
Figure 1 depicts the process of the classification scheme. The classification scheme is presented in Table 4.
Systematic reviews, therefore, emphasize the importance of evaluating the quality of the study selection. Quality assessment is an essential step in gaining a comprehensive understanding of the paper’s impact on the subject.

Publication Quality Assessment for Grey Literature

The SADACO (Significance, Authority, Date, Accuracy, Coverage, and Objectivity) checklist is used to assess grey literature. The Fourth International Conference on Grey Literature, which took place in Washington, DC, in October 1999 [45], defined grey literature as “that which is produced on all levels of government, academia, business, and industry in print and electronic formats, but which is not controlled by commercial publishers”. Grey literature encompasses a range of written materials, such as theses and dissertations, which are scrutinized by subject experts during review; conference papers, which are frequently peer-reviewed or presented by individuals with specialized knowledge; and diverse reports created by professionals working in the field. The SADACO approach proposes a mechanism for assessing the quality of grey literature. Other researchers have used a similar approach [34,35,36,37,38]. The SADACO checklist consists of 19 questions, each of which is assigned a score of 1. If the sum of the scores exceeds 50%, the source’s quality will be considered acceptable; otherwise, it will be considered a failure. Figure 2 depicts the SADACO approach.
The detailed process is explained in the following Table 5.

7. Data Extraction Strategy

7.1. Primary Study Data Extraction

The objective of this study is to collect information from pertinent publications to respond to the research questions posed in the review. The data relating to the topic under consideration will be extracted from each publication.
  • Information related to the publication, including its title, authorship, journal or conference title, and other relevant details.
  • Information or data that pertains to the research question at hand.
    The information to be recorded in the data extraction form is outlined in Table 6.

7.2. Data Extraction Process

The primary reviewer is responsible for data extraction and will approach the secondary reviewer if any issues arise. After completing the extraction process, the primary reviewer will conduct an inter-reliability test. For this test, the secondary reviewer will randomly extract data from some sources, which will be compared to the primary reviewer’s extracted data. If the results are similar, the test will be considered positive. Otherwise, the primary reviewer will review the extracted data again.
Primary Reviewer: MUHAMMAD FAISAL ABRAR.
Secondary Reviewers: DR. MUHAMMAD SOHAIL KHAN.

7.3. Data Storage

The data extraction form in the word document will be stored for each selected source. The SPSS will further be used for the Analysis of the data.

8. Data Synthesis and Analysis

Although the existing SLR guidelines [46] briefly discussed the inclusion of grey literature sources in SLR studies, most SLRs published to date have not included the quality assessment criteria for grey literature in their studies. While guidelines for SLR studies, such as [47] and SM studies [47,48] may help carry out MLRs, they do not provide adequate directions on how to assess the quality of the grey literature, in particular, because grey literature sources should be evaluated differently. When presenting the MLR process, more recent works have cited previously published MLR studies such as [49] and [50]. Our guidelines cover a much broader range of MLR literature than any previous MLR study for evaluating the quality of grey literature. To summarize a lack of MLR guidelines, there are no systematic guidelines for the quality assessment of grey literature. In comparison with previous literature [11,22,35,36,37,42], this paper aims to give a set of guidelines for the assessment of the quality of grey literature. For published literature, we found some guidelines [46,47,48] for quality assessment, but we did not find any criteria and methods for the quality assessment of grey literature. This paper gives the approach SADACO through which we assess the quality of the grey literature. The Significance, Authority, Date, Accuracy, Coverage, and Objectivity (SADACO) checklist evaluates the grey literature.
The data will be synthesized by constructing a single summary table with the columns (S.No, Credibility Factors, Frequency/Occurrences, Percentages) at the top of all credibility factors and practices with their frequencies and percentages. The data extraction form has been designed and reviewed by the authors. Most of the data are extracted from the published literature and analyzed in the following sections. The MLR protocol is currently in the implementation phase. We have got results for some of the areas mentioned earlier in the protocol. After applying the search strategy mentioned earlier to the specified digital libraries, we found 1249 papers from the five digital libraries. The preliminary and final selection information for each digital library is given in Table 7. In order to identify relevant studies for our research, we employed a snowballing technique that involved considering the references cited in the paper as well as the references in which a selected paper was cited. By performing both backward and forward snowballing, we were able to expand our initial set of relevant papers and extract 50 potential studies. Using a tollgate technique, we carefully reviewed and evaluated these 50 studies to identify the most relevant and informative papers. After this careful selection process, we ultimately considered 25 papers that met our rigorous criteria and were the most suitable for our research. With these 25 papers selected from the snowballing technique, we expanded our set of relevant studies even further by identifying additional articles using other search methods. This led to a total of 539 studies that were ultimately chosen for the data extraction process as shown in Table 7. To make it clear which studies were identified using the snowballing technique, we labeled them as ”SB” in our paper. By carefully considering the cited references and using snowballing, we were able to identify a comprehensive set of relevant studies for our research, ensuring that our findings are grounded in the most up-to-date and authoritative research in our field.
After applying the tollgate approach to our study, we selected 514 studies. We removed the duplicate studies counted as 33 in different digital libraries. Hence these studies were omitted from the final list of papers to remove duplication. At this stage of our protocol, we performed a few analyses on the finally selected studies. The first analysis is based on the location where the study is being fulfilled. Figure 3 shows the countries where research was conducted for the studies included in our MLR study. In total, 215 studies were carried out in the USA, 29 in the UK, 27 in China, 16 in India, 15 in Malaysia, 12 in Canada, 11 in Japan, and 9 in Germany, respectively.
Figure 4 shows the continents where research was conducted for the papers included in our MLR study. 48% of the study studies were achieved in North America, 26% were undertaken in Asia, 19% were conducted in Europe, 3% were conducted in Australia, and 3% were achieved in Africa.
Figure 5 shows the study-wise analysis included in our MLR study. 74% of the considered studies were published in journals, 21% belonged to Conferences, 3% were books, and 2% were dissertations.
We have categorized the journals according to impact factor and quartile. We set a threshold of five occurrences for the studies and identified the top 15 journals from the list of all journals. Table 8 shows the details. Out of all the sources, including journals, books, conferences, and theses, we drew a graph that showed that 74% of studies were carried out in journals, 21% at conferences, and 5% in books and theses, respectively. In the future, we plan to follow this review protocol to identify the factors that ensure the credibility of digital news stories. We will rank the factors using the Analytic Hierarchy Process (AHP) method.

9. Conclusions and Future Directions

The world today is bombarded with a vast amount of information from various sources, making it challenging to distinguish between credible and unreliable information. Among these sources, news plays a critical role in shaping our understanding of current events and issues. However, the credibility of news information is equally important as that of general information, especially when news is generated from multiple sources in real-time. To ensure the reliability of news articles, it is crucial to develop systematic criteria that can be used for evaluation. This research aims to identify the factors that influence the credibility of digital news stories. It is essential to recognize that news reflects a society’s traditions, ethics, and skills. Therefore, preserving it with credible information can benefit future viewers greatly. The study focuses on digital news stories since they have become a primary source of information in the current digital age. With hundreds of articles published daily, it is challenging to distinguish between credible and less credible information. Therefore, preserving news on the Web with reliable information is imperative. The Multivocal Literature Review protocol was designed to address the issue of credibility in news stories in this study. During the protocol design, the researchers found no guidelines for the quality assessment of grey literature. Grey literature refers to literature that is not commercially published, such as reports, conference proceedings, and theses. Therefore, the researchers designed guidelines for the quality assessment of grey literature, named it SADACO, which was one of the major objectives of this research. The future direction of this study is to implement the protocol and retrieve all possible credibility factors from the mentioned sources. These credibility factors will then be validated by experts through a questionnaire survey. In the last step, the researchers will prioritize the credibility factors through a Multi-Criteria Decision Making Algorithm. This approach will ensure that only the most critical factors influencing the credibility of digital news stories are considered. The importance of credible news sources cannot be overstated, especially in today’s world, where fake news and misinformation are rampant. The role of credible news sources is to inform the public, hold leaders accountable, and promote democratic values. Therefore, any study aimed at improving the credibility of digital news sources should be encouraged and supported. The systematic evaluation of news articles will also help to build trust between news organizations and their readers, further strengthening the credibility of news sources. In conclusion, the credibility of news information is of utmost importance, and systematic criteria for evaluation must be established to ensure its reliability. This research aims to identify the factors that influence the credibility of digital news stories and prioritize them using a Multi-Criteria Decision Making Algorithm. The implementation of the study’s findings will help preserve news on the Web with reliable information, making it easier for readers to distinguish between credible and unreliable news sources.

10. Validation of the Review Protocol

Our collaborator, Dr. Faisal Nadeem, reviewed the protocol at Informatics Complex, H-8, Islamabad.

Author Contributions

Conceptualization, M.F.A. and M.S.K.; methodology, M.F.A. and I.K.; software, G.A.; validation, M.F.A., M.S.K. and S.S.; formal analysis, M.F.A.; investigation, S.S.; resources, G.A.; data curation, M.S.K.; writing—original draft preparation, M.F.A.; writing—review and editing, M.F.A. and M.S.K.; supervision, M.S.K.; project administration, M.S.K.; funding acquisition, S.S. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to acknowledge Prince Sultan University and EIAS Lab for their valuable support. Further, the authors would like to acknowledge Prince Sultan University for paying the Article Processing Charges (APC) of this publication.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the authors.

Conflicts of Interest

The authors have no conflict of interest regarding the publication of this paper.

References

  1. de Kunder, M. WorldWideWebSize. 2020. Available online: http://www.worldwideWebsize.com/ (accessed on 11 March 2020).
  2. Lee, C.A. Open archival information system (OAIS) reference model. In Encyclopedia of Library and Information Sciences; Taylor & Francis: Abingdon, UK, 2010; Volume 3, pp. 4020–4030. [Google Scholar]
  3. Khan, M.; Rahman, A.U. A systematic approach towards Web preservation. Inf. Technol. Libr. 2019, 38, 71–90. [Google Scholar] [CrossRef]
  4. Khan, M. Using Text Processing Techniques for Linking News Stories for Digital Preservation. Ph.D. Thesis, Faculty of Computer Science, Preston University, Kohat, Pakistan, 2018. [Google Scholar]
  5. Khan, M.; Ur Rahman, A.; Daud Awan, M. Term-based approach for linking digital news stories. In Proceedings of the Italian Research Conference on Digital Libraries, Udine, Italy, 25–26 January 2018; pp. 127–138. [Google Scholar]
  6. Khan, M.; Rahman, A.U.; Awan, M.D.; Alam, S.M. Normalizing digital news-stories for preservation. In Proceedings of the 2016 Eleventh International Conference on Digital Information Management (ICDIM), Porto, Portugal, 19–21 September 2016; pp. 85–90. [Google Scholar]
  7. Khan, M.; Rahman, A.U. Digital news story preservation framework. In Proceedings of the of Digital Libraries: Providing Quality Information: 17th International Conference on Asia-Pacific Digital Libraries, ICADL, Seoul, Republic of Korea, 9–12 December 2015; p. 350. [Google Scholar]
  8. Hawkins, D.T. What Is Credible Information. Online 1999, 23, 86–89. [Google Scholar]
  9. Fogg, B.J.; Marshall, J.; Laraki, O.; Osipovich, A.; Varma, C.; Fang, N.; Paul, J.; Rangnekar, A.; Shon, J.; Swani, P.; et al. What makes Web sites credible? A report on a large quantitative study. In Proceedings of the of the SIGCHI Conference on Human Factors in Computing Systems, Seattle, DC, USA, 1 March 2001; pp. 61–68. [Google Scholar]
  10. Gupta, A.; Lamba, H.; Kumaraguru, P.; Joshi, A. Faking sandy: Characterizing and identifying fake images on twitter during hurricane sandy. In Proceedings of the of the 22nd International Conference on World Wide Web, Rio de Janeiro, Brazil, 13–17 May 2013; pp. 729–736. [Google Scholar]
  11. Viviani, M.; Pasi, G. Credibility in social media: Opinions, news, and health information—A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2017, 7, e1209. [Google Scholar]
  12. Rapoza, K. Can ‘Fake News’ Impact The Stock Market? Forbes News, 11 February 2017. [Google Scholar]
  13. Mendoza, M.; Poblete, B.; Castillo, C. Twitter under crisis: Can we trust what we RT? In Proceedings of the First Workshop on Social Media Analytics, Washington, DC, USA, 25–28 July 2010; pp. 71–79. [Google Scholar]
  14. Starbird, K.; Maddock, J.; Orand, M.; Achterman, P.; Mason, R.M. Rumors, false flags, and digital vigilantes: Misinformation on twitter after the 2013 boston marathon bombing. In IConference 2014 Proceedings; iSchools: Berlin, Germany, 2014; Available online: https://www.ideals.illinois.edu/items/47268 (accessed on 19 March 2021).
  15. Al-Matouq, H.; Mahmood, S.; Alshayeb, M.; Niazi, M. A Maturity Model for Secure Software Design: A Multivocal Study. IEEE Access 2020, 8, 215758–215776. [Google Scholar] [CrossRef]
  16. Neto, G.T.G.; Santos, W.B.; Endo, P.T.; Fagundes, R.A. Multivocal literature reviews in software engineering: Preliminary findings from a tertiary study. In Proceedings of the 2019 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM), Porto de Galinhas, Brazil, 19–20 September 2019; pp. 1–6. [Google Scholar]
  17. Garousi, V.; Felderer, M.; Mäntylä, M.V. Guidelines for including grey literature and conducting multivocal literature reviews in software engineering. Inf. Softw. Technol. 2019, 106, 101–121. [Google Scholar]
  18. Kapoun, J. Teaching undergrads WEB evaluation. Coll. Res. Libr. News 1998, 59, 522–523. [Google Scholar] [CrossRef] [Green Version]
  19. Jadad, A.R.; Gagliardi, A. Rating health information on the Internet: Navigating to knowledge or to Babel? JAMA 1998, 279, 611–614. [Google Scholar] [CrossRef] [PubMed]
  20. Maloney, J.C. Is advertising believability really important? Mark. Manag. 1994, 3, 47. [Google Scholar] [CrossRef]
  21. Warwick, C.; Terras, M.; Galina, I.; Huntington, P.; Pappa, N. Library and information resources and users of digital resources in the humanities. Program 2008, 42, 5–27. [Google Scholar]
  22. Fogg, B.J.; Tseng, H. The elements of computer credibility. In Proceedings of the of the SIGCHI Conference on Human Factors in Computing Systems, Pittsburgh, PA, USA, 5–20 May 1999; pp. 80–87. [Google Scholar]
  23. Fogg, B.J.; Soohoo, C.; Danielson, D.R.; Marable, L.; Stanford, J.; Tauber, E.R. How do users evaluate the credibility of Web sites? A study with over 2500 participants. In Proceedings of the 2003 Conference on Designing for User Experiences, San Francisco, CA, USA, 6–7 June 2003; pp. 1–15. [Google Scholar]
  24. Hovland, C.I.; Weiss, W. The influence of source credibility on communication effectiveness. Public Opin. Q. 1951, 15, 635–650. [Google Scholar]
  25. Iding, M.K.; Crosby, M.E.; Auernheimer, B.; Barbara Klemm, E. Web site credibility: Why do people believe what they believe? Instr. Sci. 2009, 37, 43–63. [Google Scholar] [CrossRef]
  26. Stanford, J.; Tauber, E.R.; Fogg, B.; Marable, L. Experts vs. Online Consumers: A Comparative Credibility Study Of Health and Finance Web Sites; Consumer Web Watch: Stanford, CA, USA, 2002. [Google Scholar]
  27. Cheskin, R. Trust in the wired Americas. In Cheskin Research Report; OD Practioner: Redwood City, CA, USA, 2000; pp. 1–32. [Google Scholar]
  28. Beltramini, R.F.; Stafford, E.R. Comprehension and perceived believability of seals of approval information in advertising. J. Advert. 1993, 22, 3–13. [Google Scholar] [CrossRef]
  29. Austin, E.W.; Dong, Q. Source vs. content effects on judgments of news believability. J. Q. 1994, 71, 973–983. [Google Scholar]
  30. Wathen, C.N.; Burkell, J. Believe it or not: Factors influencing credibility on the Web. J. Am. Soc. Inf. Sci. Technol. 2002, 53, 134–144. [Google Scholar] [CrossRef] [Green Version]
  31. Metzger, M.J.; Flanagin, A.J. Credibility and trust of information in online environments: The use of cognitive heuristics. J. Pragmat. 2013, 59, 210–220. [Google Scholar] [CrossRef]
  32. Mitchel, A. State of The News Media 2015. Available online: http://www.journalism.org/2015/04/29/state-of-the-news-media-2015 (accessed on 22 April 2022).
  33. Rubin, V.L.; Conroy, N.J.; Chen, Y. Towards news verification: Deception detection methods for news discourse. In Proceedings of the Hawaii International Conference on System Sciences, Kauai, Hawaii, 5–8 January 2015; pp. 5–8. Available online: https://sites.socsci.uci.edu/~lpearl/courses/readings/RubinEtAl2015_DeceptionDetectionNews.pdf (accessed on 1 March 2023).
  34. Akbar, M.A.; Mahmood, S.; Meshram, C.; Alsanad, A.; Gumaei, A.; AlQahtani, S.A. Barriers of managing cloud outsource software development projects: A multivocal study. Multimed. Tools Appl. 2022, 81, 1–24. [Google Scholar] [CrossRef]
  35. Antil, P. Requirements engineering in global scaled agile software development environment: A multi-vocal literature review protocol. arXiv 2020, arXiv:2004.12647. [Google Scholar]
  36. Lohrasbinasab, I.; Acharya, P.B.; Colomo-Palacios, R. BizDevOps: A Multivocal Literature Review. In Proceedings of the International Conference on Computational Science and Its Applications, Cagliari, Italy, 1–4 July 2020; pp. 698–713. [Google Scholar]
  37. Pereira-Vale, A.; Fernandez, E.B.; Monge, R.; Astudillo, H.; Márquez, G. Security in microservice-based systems: A multivocal literature review. Comput. Secur. 2021, 103, 102200. [Google Scholar] [CrossRef]
  38. Nylund, A. A Multivocal Literature Review on Developer Experience. Dissertation Thesis, School of Science, Aalto University, Helsinki, Finland, 2020. Available online: https://aaltodoc.aalto.fi/handle/123456789/43567 (accessed on 22 May 2022).
  39. Khan, K.; Ali, I. Data Plane Failure and Its Recovery Techniques in SDN: A Systematic Literature Review. J. King Saud Univ. Comput. Inf. Sci. 2023, 35, 176–201. [Google Scholar]
  40. Sorooshian, S.; Azizan, N.A.B. Expedited Analytical Hierarchical Process for Multicriteria Decision Making. ICIC Express Lett. 2022, 16, 145–151. [Google Scholar]
  41. Langville, A.N.; Meyer, C.D. Google’s PageRank and beyond. In Google’s PageRank and Beyond; Princeton University Press: Princeton, NJ, USA, 2011. [Google Scholar]
  42. Afzal, W.; Torkar, R.; Feldt, R. A systematic review of search-based testing for non-functional system properties. Inf. Softw. Technol. 2009, 51, 957–976. [Google Scholar] [CrossRef]
  43. Khan, R.A.; Khan, S.U.; Khan, H.U.; Ilyas, M. Systematic mapping study on security approaches in secure software engineering. IEEE Access 2021, 9, 19139–19160. [Google Scholar] [CrossRef]
  44. Alqmase, M.; Alshayeb, M.; Ghouti, L. Quality assessment framework to rank software projects. Autom. Softw. Eng. 2022, 29, 41. [Google Scholar] [CrossRef]
  45. Farace, D.J. The 4th International Conference on Grey Literature, Washington, DC, USA, 4–5 October 1999. Available online: https://cds.cern.ch/record/403957?ln=en (accessed on 3 March 2022).
  46. Shaffril, H.A.M.; Samah, A.A.; Samsuddin, S.F. Guidelines for developing a systematic literature review for studies related to climate change adaptation. Environ. Sci. Pollut. Res. 2021, 28, 22265–22277. [Google Scholar] [CrossRef] [PubMed]
  47. Kitchenham, B.; Brereton, O.P.; Budgen, D.; Turner, M.; Bailey, J.; Linkman, S. Systematic literature reviews in software engineering—A systematic literature review. Inf. Softw. Technol. 2009, 51, 7–15. [Google Scholar] [CrossRef]
  48. Wohlin, C. Guidelines for snowballing in systematic literature studies and a replication in software engineering. In Proceedings of the 18th International Conference on Evaluation and Assessment in Software Engineering, London, UK, 13–14 May 2014; pp. 1–10. [Google Scholar]
  49. Petersen, K.; Vakkalanka, S.; Kuzniarz, L. Guidelines for conducting systematic mapping studies in software engineering: An update. Inf. Softw. Technol. 2015, 64, 1–18. [Google Scholar] [CrossRef]
  50. Petersen, K.; Feldt, R.; Mujtaba, S.; Mattsson, M. Systematic mapping studies in software engineering. In Proceedings of the 12th International Conference on Evaluation and Assessment in Software Engineering (EASE) 12, Swindon, UK, 26–27 June 2008; pp. 1–10. [Google Scholar]
Figure 1. Published literature quality assessment classification scheme.
Figure 1. Published literature quality assessment classification scheme.
Applsci 13 04483 g001
Figure 2. Grey literature quality assessment process (SADACO).
Figure 2. Grey literature quality assessment process (SADACO).
Applsci 13 04483 g002
Figure 3. Country Wise Analysis.
Figure 3. Country Wise Analysis.
Applsci 13 04483 g003
Figure 4. Continent Wise Occurrence in the Literature.
Figure 4. Continent Wise Occurrence in the Literature.
Applsci 13 04483 g004
Figure 5. Study Wise Analysis.
Figure 5. Study Wise Analysis.
Applsci 13 04483 g005
Table 1. Search Strings and Databases.
Table 1. Search Strings and Databases.
S.NODatabaseSearch String
1GOOGLE SCHOLAR(“Information Credibility” OR “information believability” OR “news credibility”) AND
(“News Stories” OR “Digital News archives”) AND (“credibility factors” OR “credibility
indicators” OR “Practices” OR “Solutions”)
2ACM[[All: “information credibility”] OR [All: “digital information believability”] OR
[All: “information believability”] OR [All: “news credibility”] OR [All: “news
believability”]] AND [[All: “digital news stories”] OR [All: “news stories”] OR [All:
“digital news archives”]] AND [[All: “credibility factors”] OR [All: “credibility
indicators”] OR [All: “success factors”] OR [All: “practices”] OR [All: “solutions”]]
3IEEE XPLORE(“All Metadata”: “Information Credibility”) OR (“All Metadata”: “information
believability”) OR (“All Metadata”: “news credibility”) AND (“All Metadata”:
“News Stories”) OR (“All Metadata”: “News archives”) AND (“All Metadata”
: “credibility factors”) OR (“All Metadata”: “credibility indicators”) OR (“All Metadata”:
“information credibility Practices”)
4SCIENCE DIRECT(“Information Credibility” OR “news credibility” OR “news believability”)
AND (“News Stories” OR “Digital News archives”) AND (“credibility factors”
OR “credibility indicators” OR “Practices” OR “Solutions”)
5SPRINGER LINK(“Information Credibility” OR “information believability” OR “news credibility”)
AND (“credibility factors” OR “credibility indicators” OR “Practices” OR “Solutions”)
Table 2. MLR Search Results Record table.
Table 2. MLR Search Results Record table.
S.NODIGITAL LIBRARYInitial
Search
Title/Abstract
Based Selection
Introduction/
Discussion
Based Selection
Full Text Based
Selection
1.ACM----
2.IEEE XPLORE----
3.SCIENCE DIRECT----
4.GOOGLE SCHOLAR (Search Engine)----
5.SPRINGER LINK- --
----
TOTAL RESULTS----
Table 3. Search Result Management Format.
Table 3. Search Result Management Format.
S.NOSOURCE NAMEPAPER IDPAPER TITLE
1IEEEXPLOREIEEE-01Preliminary Analysis On The Indicators Affecting Islamic Information
Credibility In Social Media
2SCIENCEDIRECTSD-15Fake News and its Credibility Evaluation by Dynamic Relational Networks:
A Bottom up Approach
3SPRINGER LINKSL-03Credibility-Based Fake News Detection
Table 4. Classification Scheme.
Table 4. Classification Scheme.
Paper-IDPublication ChannelPublication YearResearch MethodQuality Assessment
ABCDScore
1.Conference20xxCase Study0.510.50
2.Journal20xxExperiment0111
3.Symposium20xxSLR010.50.5
4.Workshop20xxOrdinary Literature Review1.5111
5.--------
6.--------
Table 5. SADACO and relevant questions.
Table 5. SADACO and relevant questions.
Significance1. Does it enrich or add something unique, meaningful/ informative to the research context?
2. Would the research area be lesser without it?
3. Is it integral and representative and influences the work/behaviour of others?
Accuracy4. Does the item have a clearly stated aim or brief? If so, is this met?
5. Does it have a stated methodology? If so, is it adhered to?
6. Has it been peer-reviewed?
7. Supported by authoritative, documented references or credible sources?
8. If the item is secondary material (e.g., a policy brief of a technical report), refer tothe original.
Date9. Does the item have a clearly stated date related to the content?
10. Check the bibliography: have key contemporary material been included?
Authority11. Associated with a reputable organization?
12. Professional qualifications or considerable experience?
13. Cited by others? (use Google Scholar as a quick check)
14. Higher degree student under “expert” supervision?
15. Is the organization reputable? (e.g., Microsoft Corporation, IBM, PSEB etc.)
16. Is the organization an authority in the field?
Coverage17. Are any limits clearly stated?
Objectivity18. It is important to identify bias, particularly if it is unstated or unacknowledged.
19. Opinion, expert or otherwise, is still opinion: is the author’s standpoint clear?
20. Does the work seem to be balanced in presentation?
Table 6. Data extraction form.
Table 6. Data extraction form.
Date of review
Title
Authors
Database/search engine for grey literature
Methodology (interview, case study, report, survey, etc.)
Sample Population
Target Population
Publication Quality Description
Organization Type (Media Cell, News Agency, research institute etc.)
Company size (small, medium, large)
Country/location of the Analysis
Year
Critical Success Factors (CSFs)/Motivators/ Factors for ensuring credibility in digital news stories
Solutions of the CSFs for ensuring credibility in digital news stories.
News Type: political, sports, crime, cultural, entertainment, accidents, economy, science & technology etc.
Domestic/Foreign
Medium: tv, blogs, video, YouTube, social networks, etc.
Table 7. DataBases Search Results.
Table 7. DataBases Search Results.
S.NODigital LibraryInitial SearchTitle/AbstractIntroduction & DiscussionFull Text
1.ACM25242424
2.IEEE XPLORE87776159
3.SCIENCE DIRECT34302828
4.GOOGLE SCHOLAR (Search Engine)1030705385 + 113 + 31385
5.         SPRINGER LINK732914 + 214
TOTAL RESULTS1249865658514 + 25
Table 8. Quartile Based Analysis.
Table 8. Quartile Based Analysis.
S.NOJournal NameFreqTotal CitesJournal
Impact
Factor
Eigenfactor
Score
Quartile
1COMPUTERS IN HUMAN BEHAVIOR1045,0356.8290.05973Q1
2EXPERT SYSTEMS WITH APPLICATIONS555,4446.9540.04053Q1
3IEEE ACCESS6105,9683.3670.15396Q1
4MASS COMMUNICATION AND SOCIETY72.3123.3090.00334Q1
5AMERICAN BEHAVIORAL SCIENTIST57.9622.5580.00725Q1
6JOURNALISM PRACTICE182.4262.5370.0038Q1
7DIGITAL JOURNALISM142.9987.9860.00793Q1
8JOURNAL OF COMPUTER-MEDIATED
COMMUNICATION
96.4275.410.00549Q1
9JOURNALISM STUDIES113.8923.7410.00658Q1
10NEW MEDIA SOCIETY810,3258.0610.02192Q1
11JOURNALISM73.8514.4360.00607Q1
12JOURNAL OF APPLIED COMMUNICATION
RESEARCH
71.5262.8550.00126Q1
13JOURNALISM MASS COMMUNICATION
QUARTERLY
53.4354.1280.00427Q1
14JOURNAL OF COMMUNICATION811,4377.270.00963Q1
15INTERNATIONAL JOURNAL OF COMMUNICATION54.2821.8020.01176Q1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Abrar, M.F.; Khan, M.S.; Khan, I.; Ali, G.; Shah, S. Digital Information Credibility: Towards a Set of Guidelines for Quality Assessment of Grey Literature in Multivocal Literature Review. Appl. Sci. 2023, 13, 4483. https://doi.org/10.3390/app13074483

AMA Style

Abrar MF, Khan MS, Khan I, Ali G, Shah S. Digital Information Credibility: Towards a Set of Guidelines for Quality Assessment of Grey Literature in Multivocal Literature Review. Applied Sciences. 2023; 13(7):4483. https://doi.org/10.3390/app13074483

Chicago/Turabian Style

Abrar, Muhammad Faisal, Muhammad Sohail Khan, Inayat Khan, Gauhar Ali, and Sajid Shah. 2023. "Digital Information Credibility: Towards a Set of Guidelines for Quality Assessment of Grey Literature in Multivocal Literature Review" Applied Sciences 13, no. 7: 4483. https://doi.org/10.3390/app13074483

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop