Next Article in Journal
Technology Enhanced Learning Using Humanoid Robots
Previous Article in Journal
A Perfect Match: Converging and Automating Privacy and Security Impact Assessment On-the-Fly
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Language Bias in the Google Scholar Ranking Algorithm

1
Department of Communication, Universitat Pompeu Fabra, 08002 Barcelona, Spain
2
UPF Barcelona School of Management, Balmes, 134, 08008 Barcelona, Spain
*
Author to whom correspondence should be addressed.
Future Internet 2021, 13(2), 31; https://doi.org/10.3390/fi13020031
Submission received: 30 December 2020 / Revised: 17 January 2021 / Accepted: 23 January 2021 / Published: 27 January 2021
(This article belongs to the Special Issue The Current State of Search Engines and Search Engine Optimization)

Abstract

:
The visibility of academic articles or conference papers depends on their being easily found in academic search engines, above all in Google Scholar. To enhance this visibility, search engine optimization (SEO) has been applied in recent years to academic search engines in order to optimize documents and, thereby, ensure they are better ranked in search pages (i.e., academic search engine optimization or ASEO). To achieve this degree of optimization, we first need to further our understanding of Google Scholar’s relevance ranking algorithm, so that, based on this knowledge, we can highlight or improve those characteristics that academic documents already present and which are taken into account by the algorithm. This study seeks to advance our knowledge in this line of research by determining whether the language in which a document is published is a positioning factor in the Google Scholar relevance ranking algorithm. Here, we employ a reverse engineering research methodology based on a statistical analysis that uses Spearman’s correlation coefficient. The results obtained point to a bias in multilingual searches conducted in Google Scholar with documents published in languages other than in English being systematically relegated to positions that make them virtually invisible. This finding has important repercussions, both for conducting searches and for optimizing positioning in Google Scholar, being especially critical for articles on subjects that are expressed in the same way in English and other languages, the case, for example, of trademarks, chemical compounds, industrial products, acronyms, drugs, diseases, etc.

1. Introduction

A researcher’s professional career is heavily dependent on the visibility of, and the recognition afforded by, their scholarly output. The number of citations received and the corresponding indexes associated with this variable, in particular the h-index, are typically the most widely used metrics employed in official processes of accreditation before appropriate academic boards or commissions. Indeed, the need to be cited, combined with the exponential increase in world bibliographic output, means that researchers today have to promote their own articles as the final step in the complex process of publishing their research findings. This promotion of their research output usually also implies building their personal academic brand [1,2], including the creation of complementary content that extends well beyond traditional scholarly articles for specialized publications or papers for conferences.
Among the actions researchers can take to promote both their personal brand and scientific output are mentioning their articles in academic and non-academic social networks, creating a professional blog with complementary content (videos, presentations, PDFs), creating profiles on a range of platforms from ORCID, Google Scholar Citation, and ResearcherID to Mendeley, depositing documents in open access repositories and optimizing articles so that they command a good position in search engines, especially Google Scholar. Indeed, the vast majority of Internet users do not look beyond the second or third page of results [3], which means for a document to be found easily it has to be optimized to ensure it appears towards the top of the first page.
Search results are ranked according to relevance, a value automatically calculated by search engines. Moreover, this ranking is usually established as the default order, because other forms of sorting—for example, by title or by date—are considered less significant for most search intentions (although these other forms of sorting are often available). Relevance is calculated using an algorithm that takes into account a range of factors, which means that for each search engine, relevance will differ to some degree, given that it is being defined by a distinct algorithm in each case.
Search engine optimization (SEO) [4,5] is, today, a well-established discipline, the goal of which is to highlight the quality of web pages and so improve their position in the results pages. This goal should not be achieved by fraudulent means, but depends rather on knowing how the algorithms that determine relevance operate, identifying the factors taken into account and, finally, optimizing these factors in one’s documents. However, Google Scholar is not always able to detect attempts at manipulation [6,7].
Today, there is a huge community of SEO experts and companies that dedicate their efforts to analyzing and discussing Google’s relevance ranking algorithm. Via blogs [8,9,10,11], online publications [12,13,14], and books [15,16,17] they advise designers and webmasters as to how they can optimize their websites so that they are easily indexed and can occupy the highest rankings in the results pages.
Google’s relevance algorithm is based on more than 200 factors [18] and include the number of links received, the keywords and related terms in the title and other significant areas of the document, the download speed of the server on which the page is hosted, the length of the text, the user experience, the mobile-first design, the semantic tagging, the age of the domain, etc. Google has never released complete information about all these factors or the exact weighting attached to each; the company only provides general, incomplete information in order to avoid spam. Indeed, if all the details of how the algorithm works were known, then poor quality documents could be placed at the top of the results page.
This “black box” policy has led SEO professionals to conduct reverse engineering research in an effort to identify the specific factors involved in relevance ranking. Thus, they analyze the search results in order to infer how the algorithm works. However, it is a complicated process in which many factors intervene and it is not easy to draw any conclusive results.
In recent years, this ecosystem of research concerned with algorithms and the subsequent publication of recommendations has been extended to Google Scholar and academic articles. On a much smaller scale, reverse engineering research has been applied to Google Scholar [19,20,21,22,23,24,25] and blogs [26,27,28,29,30], university library guidelines [31,32,33,34] and the authors’ services of the publishers of academic journals [35,36,37,38,39,40] offer their recommendations as to how to optimize articles so that they appear at the top of the rankings of Google Scholar’s results pages. This SEO applied to academic search engines has been called academic search engine optimization or ASEO [20,22,41,42,43].
This research community is still in its infancy both in terms of the quantity and quality of its output; moreover, the recommendations given for Google Scholar are often contaminated by research findings for Google. In fact, they are two quite distinct algorithms that operate on two quite distinct types of document in two very different environments. Indeed, as far as the ranking algorithm is concerned, academic documents have at least four major characteristics that clearly distinguish them from web pages: most are in PDF (and not HTML) format; they contain links based on bibliographic citations with other academic documents (not hyperlinks); once published, they are not modified; and, usually, author metadata and the date of publication are clearly identified.
Promoting a personal academic brand and the visibility of web pages, blogs, videos and other complementary content depend to a large extent on Google positioning. But the visibility of academic articles or conference papers is determined by their optimization for Google Scholar. These differences need to be clarified, while it is necessary to further our understanding of Google Scholar’s relevance ranking algorithm, which is not so well known and widely analyzed as Google’s general search algorithm.
The aim of this study is to do just that and, more specifically, here, because of its far-reaching implications, we seek to determine whether the language in which a document is written is a key positioning factor. In this regard, no previous study, to the best of our knowledge, has attempted to find a relationship between positioning and language, be it for Google Scholar or for the general Google search engine.
Normally, language plays no role as a ranking factor in keyword searches, given that the language of the search word itself determines the language of the documents retrieved. If the documents are written in the same language, this factor is overridden. Language only intervenes in those few cases of keywords with the same spelling in different languages that generate multilingual lists of results. In contrast, searches by author or year are conducted independently of language and always provide what we shall refer to henceforth as multilingual results (or searches).
When searches are multilingual, that is, when results are provided in different languages for the same search, the language of the documents can be a decisive factor if it can be shown that this conditions ranking. Thus, our primary research question here is the following: In multilingual searches, is the language in which a document is written a factor in Google Scholar’s ranking algorithm?
Our hypothesis is that Google Scholar favors the English language in multilingual search results. As a result, documents in other languages have fewer possibilities of being placed at the top of the rankings for the sole reason that they are not published in English.
In the following section, we discuss related studies in the literature, before moving on to present the applied research methodology and the method used to select our sample. Next, we analyze the results obtained from our statistical data and from observation of our scatter plots. The limitations of the study are discussed and new lines of research are proposed. Finally, in the conclusions, the repercussions of our findings are highlighted, both for searches and for the optimization of positioning in Google Scholar.

2. Prior Studies

Google Scholar has been the subject of numerous studies, aimed, above all, at assessing its quality [44,45,46,47,48,49,50,51,52,53] and search effectiveness [45,54], determining whether it is a suitable tool for conducting bibliometric studies [55,56,57,58,59,60,61] and evaluating author impact using the h-index [25,56,62,63,64,65,66,67].
In contrast, few studies have specifically examined how the Google Scholar ranking algorithm works [19,20,21,22]. One of the reasons for this is probably the erroneous belief that Google Scholar uses the same or a very similar algorithm to the one used by the general Google search engine. However, the few studies of ranking by relevance in Google Scholar that have been conducted [19,20,21,22,23,24,25] conclude that it is a specific algorithm, adapted to the particular characteristics of scholarly documents; nevertheless, some ranking factors may be the same in both cases, such as the weight attached to keywords in the title [19]. Other factors, though, are clearly specific, such as the number of citations received, which is unique to Google Scholar [21,25].
Moreover, the frequency of keywords in the text of the article does not appear to be a factor that the Google Scholar algorithm takes into account [19]. Yet, there is evidence that the date of publication is of relevance, given that according to Beel and Gipp [19] older articles have a higher ranking than more recent publications.
A further factor identified by various authors is that the search term has to coincide exactly with a term in the documents retrieved and, moreover, that Google Scholar does not expand its searches through synonyms of keywords in the same way that Google does [68,69].
Where there is greater consensus is in attributing considerable weight to the number of citations received in Google Scholar’s calculation of ranking by relevance [19,23,24,25]. This is hardly surprising, given that Google Scholar is a bibliographic database and traditionally tools of this type use citations received as a means for sorting their results page.

3. Methodology

As indicated, the aim of this study is to further our understanding of the Google Scholar relevance ranking algorithm. We are particularly interested in determining whether the language in which a document is written affects its position on the results page, in other words, if language is one of the factors that intervenes in this algorithm when results exist in several languages.
We have employed reverse engineering techniques as our research methodology. This method is frequently used to determine how a device works by undertaking an analysis of its behavior and results. The method is also used to extract the source code from the compiled files.
Reverse engineering has been used for years to study search engine ranking algorithms [70,71,72]. The documents that occupy the highest rankings in the results lists are analyzed and, then, by means of a statistical analysis it can be deduced whether the factors they present intervene in the ranking. The most frequently used statistical test in this context is Spearman’s correlation, given that the data never form normal distributions. The native search engine ranking is compared with the rankings created by the researcher using some of the frequent characteristics presented by the highest-ranking documents. The higher the correlation coefficient, the more similar the two rankings are and, therefore, the greater the weight that can be attributed to the alternative ranking factor in the search engine algorithm under analysis.
In the case of Google’s algorithm, more than 200 factors are considered [73,74]. This algorithm is highly complex and, moreover, it is subject to constant improvements, a process in which the application of artificial intelligence is involved [75]. It is for this reason that in studies of Google correlation coefficients rarely exceed 0.3. These results are very low, but sufficient to obtain some indication as to whether a particular factor under analysis is included in the computation of the ranking of documents.
Google Scholar adheres to the same (dis)information policy as Google and does not provide specific details as to how the algorithm works, officially, in an effort to prevent spam. This means it publishes only very general guidelines:
“Google Scholar aims to rank documents the way researchers do, weighing the full text of each document, where it was published, who it was written by as well as how often and how recently it has been cited in other scholarly literature.”
[76]
However, thanks to the findings of previous studies [57,77], it is evident that the Google Scholar algorithm is simpler than that of the general Google search engine. This has certain advantages for reverse engineering studies of Google Scholar, with correlation coefficients higher than 0.8 often being obtained, which allows more reliable conclusions to be drawn.
The sample data for this study consisted of a total of 45 searches each producing 1000 results, which means we analyzed approximately 45,000 items of information, a similar number to other studies of the same type in which reverse engineering was employed [70,71,78] to study the Google Scholar algorithm [19,20,21,22,23,28].
More specifically, we conducted three different types of search, obtaining 15,000 results in each case: searches by author, by year, and by keyword. By so doing, our aim was to analyze the algorithm in different contexts and to identify similar patterns of behavior that would lend more robustness to our results.
In the case of searches by keyword, we selected the most frequent terms meeting the following conditions:
  • General, non-specialized, language terms, in order to avoid any thematic biases.
  • Terms spelt the same in English and Spanish (i.e., cognates), in order to generate multilingual searches in at least these two languages. Subsequently, we also detected results in Portuguese and other languages, as the spelling of some of the terms selected also coincided with the spellings in these languages.
Additionally, by using the search form, we forced the appearance of the keyword in the title of the documents found so that, in this way, we could neutralize this factor and prevent it from distorting our results.
In the case of searches by year, we included documents published before 2015, in order to avoid more recent documents that might not have had sufficient time to generate a significant number of citations. As we report below, citations are a central element in this methodology.
Finally, in the case of searches by author, to generate as many multilingual results as possible, we selected the most frequently used surnames of Hispanic origin in the United States. Note, surnames were used in isolation from any first names.
To avoid data bias due to low data volumes, the following conditions were also applied to all searches in the sample:
  • At least 2.5% of the documents in all the searches conducted had to be written in languages other than English.
  • All documents included in the analysis had to have at least 15 citations received. The reason for our introducing this condition was to avoid a high number of documents with zero citations received which would have made it impossible to differentiate their ranking by citations.
Our final sample was made up as follows:
  • Keywords: capital, combustible, crisis, cultural, federal, festival, final, general, idea, invisible, moral, popular, social, terror, total.
  • Authors: Cruz, Diaz, Flores, Garcia, Gomez, Gutierrez, Hernandez, Martinez, Morales, Ortiz, Perez, Ramirez, Ramos, Sanchez, Torres.
  • Years: from 2000 to 2014.
Table 1 shows the percentages of documents written in languages other than English retrieved in each search. The data for this study were collected between 28 August 2020 and 22 September 2020 using the Publish or Perish tool [79,80]. Searches by keyword and author were performed without accents and in lower case. The language of each item was identified by employing Google’s language recognition technology by primarily analyzing the title of the documents. This technology is available among the spreadsheet functions of Google Drive.
For the statistical study, as the distributions were not normal according to the Kolmogorov-Smirnov test, we used Spearman’s correlation.
For each of the searches analyzed, the number of citations received was transformed to an ordinal scale. In this way, we constructed an alternative ranking that could subsequently be correlated with the native Google Scholar ranking. This is a very common procedure [19,21] in this type of research as it allows the comparison to be more naturally made between two ordinal variables.
As mentioned, the number of citations a document receives is a very important factor in the Google Scholar ranking algorithm, especially in searches by date and author, when keywords do not intervene. In searches of this type, the ranking by number of citations received is very similar to the native Google Scholar ranking, with correlation coefficients higher than 0.9 [23].
To determine the role that language plays in Google Scholar’s ranking, we exploited this similarity. Our aim in so doing was to determine whether documents written in English follow a similar pattern to documents written in other languages when the native ranking is compared with the ranking by citations received. We wished to see if the number of citations received is an equally determining factor regardless of the language in which the document is written.
If the correlations of the documents written in English are similar to those of documents written in other languages, it means that language does not condition the ranking algorithm; on the other hand, if the correlations are different we will have found indications that language may play a role in this ranking, be it positive or negative.
In order to obtain global values of the correlations, we calculated the median values of the ranking by citations of the 45 searches making up the sample and for each of the 1000 document positions in the results list. These one thousand medians (for each of the 45 searches) provide a measure of central tendency of the citation ranking which can then be correlated with the native Google Scholar ranking.
To conduct the statistical analysis, we used R, version 3.4.0 [71]; SPSS, version 20.0.0; Excel, and Google Sheets. Confidence intervals were constructed using the normal approximation method and applying Fisher’s transformation using the R psych package [81,82,83]. Fisher’s transformation when applied to Spearman’s correlation coefficient is asymptotically normal. The scatter plots were created with Google Sheets and Tableau.

4. Results

In line with previous research [23,24], an analysis of our data verified a high correlation between the ranking by the number of citations received and Google Scholar’s native ranking. The high value of these coefficients indicates that a document receiving a high ranking on the basis of the number of citations received is also highly positioned in Google Scholar’s native ranking.
However, when we focused our analysis solely on documents written in languages other than English, we detected an unusual ranking pattern, one that had not previously been identified. This pattern can be defined in terms of the following two characteristics:
  • In multilingual searches in Google Scholar, documents written in languages other than English fail, in the main, to appear among the first 900 results on the search page (i.e., they present a ranking >900).
  • The Google Scholar ranking algorithm works differently for documents written in languages other than English, as it does not consider the number of citations received as the main ranking factor.
This implies that while documents written in English are awarded a good position based on the number of citations received, this same factor, even when the number of citations is high, does not benefit documents written in other languages. In other words, in multilingual searches in Google Scholar, the widely recognized positioning factor based on the number of citations received is not applicable to documents that are not published in English.
Below, we present in detail the data tables and scatter plots that corroborate these conclusions.

4.1. Data Analysis Identifying the Different Ranking Pattern Presented by Documents Written in Languages other than English

Table 2, Table 3 and Table 4 show the percentages of documents in languages other than in English which failed to appear among the top 900 search results. It is evident that they constitute the vast majority: 92.1% in the case of searches by year, 85.3% in those by author and 89.7% in those by keyword.
Table 5 provides a summary of these results, indicating that, overall, 91% of documents in languages other than English are ranked above 900 (i.e., >900). This contrasts with just 3.7% of documents in English that appear in these positions. It is also striking that only 0.7% of documents in languages other than English appear in the top 100 positions. This figure falls to 0.2% when the first 20 positions are analyzed (Table 5).
Table 6, Table 7 and Table 8 show the correlation coefficients between the ranking by the number of citations received and the Google Scholar native ranking. The results in English are differentiated from those of the other languages in order to highlight the different ranking pattern identified. In all cases, the correlations are higher for the documents published in English. These differences are especially notable in searches by years and by authors with values always higher than 0.82 in the case of documents in English and values always lower than 0.44 in documents published in other languages.
In the case of searches by keyword, some outcomes point to more moderate differences, but if we observe the correlations between the median values, the differences are again very high with values of 0.91 for documents in English and 0.05 for those in other languages, although this last value is not statistically significant (Table 8). In fact, most of the correlations of the results for documents in other languages are not statistically significant, another indication that we are dealing with two different distributions.
If we take the overall data (Table 9) for all the searches and compare the correlations of the medians of the two rankings, we obtain values of 0.97 for the documents in English and of just 0.18 for those in other languages.
The value of the correlation for all the data is, likewise, very high at 0.91. As discussed previously, this indicates that while the number of citations is a very important factor in the ranking algorithm, the low values recorded in the case of documents written in languages other than English shows that this factor is not being applied and, as a result, the behavior of the ranking algorithm is unusual in these cases.

4.2. Analysis of Scatter Plots to Show the Differ Ranking Pattern of Documents Written in Languages other than English

The different ranking pattern presented by documents written in languages other than English is also evident in the scatter plots (Figure 1, Figure 2, Figure 3 and Figure 4), where the documents published in English are represented by gray dots and those corresponding to other languages are represented by red dots. The blue dots represent the median values of the ranking by citations of the documents listed among the top thousand in the results page. As is evident, the vast majority of the results corresponding to documents written in languages other than English (red dots) do not appear in the top 900 positions. There are, in fact, two-point clouds: One essentially for documents published in English and one for documents published in other languages.
Documents in languages other than English do not adhere to the pattern set by the overall results, a pattern that places the points along the diagonal of the scatter plot, corresponding to a high correlation coefficient. A pattern that gives a good position in the ranking according to the number of citations received also gives a good position in the Google Scholar native ranking. The breakdown in this pattern is especially notable in the case of the dots located in the lower right corner of the four graphs, corresponding to documents that appear in the top 100 positions in the ranking by number of citations, but which are located above rank position 900 in the Google Scholar native ranking.
This second point cloud that forms above rank position 900 has been previously detected by Beel [19] and Rovira [23], but these authors did not attribute its origin to the language in which the documents were published.
The scatter plots illustrate quite clearly the same inconsistency in the ranking algorithm that we report above in our data analysis. This inconsistency implies that documents written in languages other than English are treated in a discriminatory manner.

5. Discussion

Our data are insufficient to be able to conclude whether the ranking of documents in languages other than English is due simply to an accidental bias, caused by an unintentional error in the software program, or whether it is an intentional outcome of the design of the algorithm, based, for example, on a belief that the English language has what might be considered an internationally superior dimension to that of other languages. This would explain (although we cannot state with any certainty) why in multilingual searches documents written in other languages become virtually invisible.
Neither do we have sufficient data to determine whether, in fact, two different algorithms are being used, one for English and one for all other languages, or whether the documents in English are ranked first and only then are the documents in all other languages ranked employing the same algorithm, to mention just a few possibilities.
However, what we have been able to demonstrate is that the Google Scholar relevance ranking algorithm favors documents in English and handicaps documents written in other languages in searches that produce multilingual results. Moreover, the vast majority of documents written in languages other than English have no chance at all of being found in searches of this type, even if they are high quality articles with hundreds or, even, thousands of citations received. Indeed, for 94% of documents of this type their possibilities of being found are absolutely zero, given that almost nobody accesses items located above a rank position of 700 on a list of search results (Table 5).
More specifically, only 0.2% of these non-English documents have a chance of being found in a multilingual search in Google Scholar, given that they would occupy a place in the top 20 rankings. However, if the ranking algorithm were to be applied homogeneously across all documents, giving the same weight to citations received in all of them, the percentage of non-English documents among these top 20 positions would be 3.2%, i.e., much higher than the value of 0.2% actually obtained (Table 5).
This anomalous behavior is detected in multilingual searches that return results in different languages, that is, searches conducted, for example, by year, by author, or by keywords with identical spelling in more than one language. Clearly, in typical keyword searches, in which all the results are in the same language, this bias does not occur.
We can, moreover, rule out the possibility that other variables, in addition to the language in which a document is written, might be intervening, to produce this anomalous ranking. We have been able to verify that these results only occur in documents written in a language other than English and at such high percentages that the possibility of the participation of any other factors can be discarded.
Despite the restrictions we imposed on our data collection, the number of results in languages other than English is relatively low compared to the whole sample of data analyzed: 4.4% in the case of keywords, 7.7% in that of authors and 8.4% in the case of years (Table 1). These low percentages are also a consequence of a biased ranking. Future studies need to include searches with a more balanced number of results between documents in English and those published in other languages in order to corroborate that the ranking is also discriminatory in these cases. However, the only way to obtain percentages close to 50% is by conducting unnatural searches, such as combining years and keywords of one or two letters.
Finally, it is worth highlighting two additional aspects that might also be the object of future research. First, we have reported that the point cloud for searches by keyword is much more dispersed than those for searches by year or by author. This dispersion is a clear indication that in searches by keyword other factors, in addition to the number of citations received, are intervening. For this reason, more research is needed to identify the other factors that contribute to this dispersion. Second, it would be of interest to determine whether the bias we report differs according to the non-English language used for publication, in other words, whether documents written in Spanish, French or German, for example, present a specific ranking pattern in Google Scholar searches.

6. Conclusions

Both standard SEO and ASEO have a primary concern with determining the main factors employed in ranking algorithms. The aim of these two fields of enquiry is to make more visible the characteristics that articles present and which, in turn, serve as criteria for ranking. This is the case of so-called “white hat SEO”, that is, positioning actions that seek to optimize the visibility of information by ethical means.
This need to identify positioning factors has motivated and guided our research here, which reports an unexpected result with major repercussions for researchers. We have demonstrated that in Google Scholar, in the case of multilingual results, documents in languages other than English are systematically relegated to positions that make them virtually invisible. These positions are not the result of applying the standard ranking algorithm, as is done for documents in English. Moreover, in practical terms, it is of little importance whether this bias is accidental or the result of an algorithm designed specifically to rank differently according to the language in which a document is written.
However, it is important to draw attention to this major bias so that it might be limited or compensated for when a search is conducted. A lack of awareness of this factor could be detrimental to researchers from all over the non-English-speaking world, making them believe that there is no literature in their national language when they conduct searches with multilingual results. This is particularly the case in the most frequent searches, that is, those conducted by year. Nevertheless, it can also occur in searches using certain keywords that are the same in languages around the world, including trademarks, chemical compounds, industrial products, acronyms, drugs, and diseases, with Covid-19 being the most recent example.
Moreover, if we consider the results of this study from the perspective of ASEO, it is more than evident that until this bias is addressed, the chances of being ranked in a multilingual Google Scholar search increase remarkably if the researchers opt for publication in English. This recommendation is especially critical for researchers working in the areas mentioned in the previous paragraph whose names are the same in English and in their national language.

Author Contributions

Conceptualization, C.R.; Methodology, C.R.; Validation L.C. and C.L.; Investigation C.R., and L.C.; Resources, C.R. and C.L.; Data Curation, C.R.; Writing—Original Draft Preparation, C.R.; Writing—Review and Editing, L.C. and C.L.; Supervision L.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is part of the Project “Interactive storytelling and digital visibility in interactive documentary and structured journalism”. RTI2018-095714-B-C21 (Micinn/Feder) Ministry of Science, Innovation and Universities of Spain.

Data Availability Statement

[dataset] Cristòfol Rovira; Lluís Codina; Carlos Lopezosa. 2021. Data set of the article: Language Bias in the Google Scholar Ranking Algorithm; Zenodo; Version 1; http://doi.org/10.5281/zenodo.4463201.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. González-Solar, L. Marca personal en entornos académicos: Una perspectiva institucional. An. Doc. 2018, 21. [Google Scholar] [CrossRef]
  2. Harzing, A.W. Building Your Academic Brand through Engagement with Social Media. 2018. Available online: https://harzing.com/blog/2018/05/building-your-academic-brand-through-engagement-with-social-media (accessed on 17 October 2020).
  3. Marcos, M.-C.; González-Caro, C. Comportamiento de los usuarios en la página de resultados de los buscadores. Un estudio basado en eye tracking. El Prof. de la Inf. 2010, 19, 348–358. [Google Scholar] [CrossRef] [Green Version]
  4. Yalçın, N.; Köse, U. What is search engine optimization: SEO? Procedia Soc. Behav. Sci. 2010, 9, 487–493. [Google Scholar] [CrossRef] [Green Version]
  5. Ziakis, C.; Vlachopoulou, M.; Kyrkoudis, T.; Karagkiozidou, M. Important factors for improving Google search rank. Future Internet 2019, 11, 32. [Google Scholar] [CrossRef] [Green Version]
  6. López-Cózar, D.E.; Robinson-García, N.; Torres-Salinas, D. Manipulating Google Scholar citations and Google Scholar metrics: Simple, easy and tempting. arXiv 2012, arXiv:1212.0638. [Google Scholar]
  7. López-Cózar, D.E.; Robinson-García, N.; Torres-Salinas, D. The Google Scholar experiment: How to index false papers and manipulate bibliometric indicators. J. Assoc. Inf. Sci. Technol. 2014, 65, 446–454. [Google Scholar] [CrossRef]
  8. SEMrush Blog. Available online: https://www.semrush.com/blog/ (accessed on 1 July 2020).
  9. The Moz Blog. Available online: https://moz.com/blog (accessed on 1 July 2020).
  10. Yoast SEO Blog. Available online: https://yoast.com/seo-blog/ (accessed on 1 July 2020).
  11. Developers Google. Available online: https://developers.google.com/search/blog#blog-principal-de-la-busqueda-de-google (accessed on 1 July 2020).
  12. Search Engine Journal. Available online: https://www.searchenginejournal.com/ (accessed on 1 July 2020).
  13. Search Engine Land. Available online: https://searchengineland.com/ (accessed on 1 July 2020).
  14. Search Engine Watch. Available online: https://www.searchenginewatch.com/ (accessed on 1 July 2020).
  15. Clarke, A. SEO 2020 Learn Search Engine Optimization with Smart Internet Marketing Strategies: Learn SEO with Smart Internet Marketing Strategies; Newest, Ed.; EEUU: Trenton, NJ, USA, 2019; pp. 1–244. [Google Scholar]
  16. Kent, P. SEO For Dummies; EEUU: Trenton, NJ, USA, 2019; pp. 1–478. [Google Scholar]
  17. Maciá, F. SEO Avanzado. Casi todo lo Que sé Sobre Posicionamiento Web; Anaya: Barcelona, Spain, 2020; pp. 1–416. [Google Scholar]
  18. Google. How Google Search Works. Learn How Google Discovers, Crawls, and Serves Web Pages. Available online: https://support.google.com/webmasters/answer/70897?hl=en (accessed on 1 July 2020).
  19. Beel, J.; Gipp, B. Google scholar’s ranking algorithm: An introductory overview. In Proceedings of the 12th International Conference on Scientometrics and Informetrics, ISSI’09, Istanbul, Turkey, 14–17 July 2009; pp. 230–241. Available online: https://goo.gl/c8a6YU (accessed on 23 January 2021).
  20. Beel, J.; Gipp, B. Google Scholar’s ranking algorithm: The impact of articles’ age (an empirical study). In Proceedings of the Sixth International Conference on Information Technology: New Generations, ITNG’09, Las Vegas, NA, USA, 27–29 April 2009; pp. 160–164. [Google Scholar] [CrossRef]
  21. Beel, J.; Gipp, B. Google scholar’s ranking algorithm: The impact of citation counts (an empirical study). In Proceedings of the Third International Conference on Research Challenges in Information Science, RCIS 2009c, Nice, France, 22–24 April 2009; pp. 439–446. [Google Scholar] [CrossRef]
  22. Beel, J.; Gipp, B.; Wilde, E. Academic search engine optimization (ASEO). Optimizing scholarly literature for Google Scholar & co. J. Sch. Publ. 2010, 41, 176–190. [Google Scholar] [CrossRef]
  23. Rovira, C.; Guerrero-Solé, F.; Codina, L. Received citations as a main SEO factor of Google Scholar results ranking. El Prof. de la Inf. 2018, 27, 559–569. [Google Scholar] [CrossRef] [Green Version]
  24. Rovira, C.; Codina, L.; Guerrero-Solé, F.; Lopezosa, C. Ranking by relevance and citation counts, a comparative study: Google Scholar, Microsoft Academic, WoS and Scopus. Future Internet 2019, 11, 202. [Google Scholar] [CrossRef] [Green Version]
  25. Martín-Martín, A.; Orduña-Malea, E.; Ayllón, J.M.; López-Cózar, E.D. Does Google Scholar contain all highly cited documents (1950–2013). arXiv 2014, arXiv:1410.8464. [Google Scholar]
  26. Sparks, A. 8 Winning Hacks to Use Google Scholar for Your Research Paper, Editage. 2018. Available online: https://www.editage.com/insights/8-winning-hacks-to-use-google-scholar-for-your-research-paper (accessed on 1 July 2020).
  27. Cordero, J.J. 8 Estrategias Para Promocionar los Artículos Profesionales de tu Blog. 2018. Available online: https://www.javiercordero.com/como-promocionar-articulos-blog/ (accessed on 1 July 2020).
  28. Florido, M. Google Académico—7 Consejos para mejorar el posicionamiento. 2015. Available online: https://www.marketingandweb.es/marketing/google-academico/ (accessed on 1 July 2020).
  29. Drew, C. 11 Best Tips on How to use Google Scholar. 2020. Available online: https://helpfulprofessor.com/google-scholar/ (accessed on 1 July 2020).
  30. Miles, S. 12 Tips for Increasing Your Visibility in Google Search. 2020. Available online: https://webpublisherpro.com/12-tips-for-increasing-your-visibility-in-google-search/ (accessed on 1 July 2020).
  31. UCA Library. Research Visibility. SEO for Authors: A How-to Guide. Available online: https://guides.library.ucla.edu/seo/author (accessed on 2 July 2020).
  32. UNED. Investiga UNED. 10 Consejos Para Difundir tu Investigación y Conseguir más Impacto. 2017. Available online: http://investigauned.uned.es/10-consejos-para-difundir-tu-investigacion-y-conseguir-mas-impacto/ (accessed on 2 July 2020).
  33. University of Pittsburgh. How to Increase the Visibility of Your Research? 2020. Available online: https://pitt.libguides.com/researchvisibility (accessed on 2 July 2020).
  34. University of Montana. Google Scholar at the University of Montana. 2020. Available online: https://libguides.lib.umt.edu/c.php?g=854135&p=6115806 (accessed on 2 July 2020).
  35. Wiley. Search Engine Optimization (SEO) for Your Article. Available online: https://authorservices.wiley.com/author-resources/Journal-Authors/Prepare/writing-for-seo.html (accessed on 2 July 2020).
  36. Emerald Publishing. How to… Make Your Research Easy to Find with SEO. Available online: https://www.emeraldgrouppublishing.com/services/authors/author-how-guides/make-your-research-easy-find-seo (accessed on 2 July 2020).
  37. Sage Publishing. Promote Your Article. Available online: https://uk.sagepub.com/en-gb/eur/promote-your-article (accessed on 2 July 2020).
  38. Plos. Spreading the Word about Your Research. Available online: https://plos.org/article-promotion/ (accessed on 2 July 2020).
  39. Taylor and Francis. Search Engine Optimization for Academic Articles. Available online: https://authorservices.taylorandfrancis.com/research-impact/search-engine-optimization-for-academic-articles/ (accessed on 2 July 2020).
  40. Elsevier. Get Found—Optimize Your Research Articles for Search Engines. 2012. Available online: https://www.elsevier.com/connect/get-found-optimize-your-research-articles-for-search-engines (accessed on 2 July 2020).
  41. Codina, L. SEO Académico: Definición, Componentes y Guía de Herramientas. 2019. Available online: https://www.lluiscodina.com/seo-academico-guia (accessed on 9 July 2020).
  42. Martín-Martín, A.; Ayllón, J.M.; Orduña-Malea, E.; López-Cózar, E.D. Google Scholar metrics released: A matter of languages and something else. arXiv 2016, arXiv:1607.06260v1. [Google Scholar]
  43. Muñoz-Martín, B. Incrementa el impacto de tus artículos y blogs: De la invisibilidad a la visibilidad. Rev. Soc. Otorrinolaringol. Castilla León Cantab. Rioja 2015, 6, 6–32. Available online: http://hdl.handle.net/10366/126907 (accessed on 23 January 2021).
  44. Giustini, D.; Boulos, M.N.K. Google Scholar is not enough to be used alone for systematic reviews. Online J. Public Health Inform. 2013, 5, 214. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Walters, W.H. Google Scholar search performance: Comparative recall and precision. Portal-Libr. Acad. 2008, 9, 5–24. [Google Scholar] [CrossRef]
  46. De-Winter, J.; Zadpoor, A.; Dodou, D. The expansion of Google Scholar versus Web of Science: A longitudinal study. Scientometrics 2014, 98, 1547–1565. [Google Scholar] [CrossRef]
  47. Harzing, A.W. A preliminary test of Google Scholar as a source for citation data: A longitudinal study of Nobel prize winners. Scientometrics 2013, 94, 1057–1075. [Google Scholar] [CrossRef]
  48. Harzing, A.W. A longitudinal study of Google Scholar coverage between 2012 and 2013. Scientometrics 2014, 98, 565–575. [Google Scholar] [CrossRef] [Green Version]
  49. De-Groote, S.L.; Raszewski, R. Coverage of Google Scholar, Scopus, and Web of Science: A case study of the h-index in nursing. Nurs Outlook 2012, 60, 391–400. [Google Scholar] [CrossRef]
  50. Orduña-Malea, E.; Ayllón, J.M.; Martín-Martín, A.; Delgado-López-Cózar, E. About the size of Google Scholar: Playing the numbers. arXiv 2014, arXiv:1407.6239. [Google Scholar]
  51. Orduña-Malea, E.; Ayllón, J.M.; Martín-Martín, A.; Delgado-López-Cózar, E. Methods for estimating the size of Google Scholar. Scientometrics 2015, 104, 931–949. [Google Scholar] [CrossRef] [Green Version]
  52. Pedersen, L.A.; Arendt, J. Decrease in free computer science papers found through Google Scholar. Online Inf. Rev. 2014, 38, 348–361. [Google Scholar] [CrossRef] [Green Version]
  53. Jamali, H.R.; Nabavi, M. Open access and sources of full-text articles in Google Scholar in different subject fields. Scientometrics 2015, 105, 1635–1651. [Google Scholar] [CrossRef]
  54. Jamali, H.R.; Asadi, S. Google and the scholar: The role of Google in scientists’ information-seeking behaviour. Online Inf. Rev. 2010, 34, 282–294. [Google Scholar] [CrossRef] [Green Version]
  55. Aguillo, I.F. Is Google Scholar useful for bibliometrics? A webometric analysis. Scientometrics 2012, 91, 343–351. [Google Scholar] [CrossRef]
  56. Jacsó, P. Calculating the h-index and other bibliometric and scientometric indicators from Google Scholar with the Publish or Perish software. Online Inf. Rev. 2009, 33, 1189–1200. [Google Scholar] [CrossRef] [Green Version]
  57. Torres-Salinas, D.; Ruiz-Pérez, R.; Delgado-López-Cózar, E. Google scholar como herramienta para la evaluación científica. El Prof. de la Inf. 2009, 18, 501–510. [Google Scholar] [CrossRef] [Green Version]
  58. Beel, J.; Gipp, B. Academic search engine spam and Google Scholar’s resilience against it. J. Electron. Publ. 2010, 13. [Google Scholar] [CrossRef]
  59. Delgado-López-Cózar, E.; Robinson-García, N.; Torres-Salinas, D. Manipular Google Scholar citations y Google Scholar metrics: Simple, sencillo y tentador. In EC3 Working Papers; Universidad De Granada: Granada, Spain, 2012; Available online: http://hdl.handle.net/10481/20469 (accessed on 1 July 2019).
  60. Meho, L.; Yang, K. Impact of Data Sources on Citation Counts and Rankings of LIS Faculty: Web of Science Versus Scopus and Google Scholar. J. Assoc. Inf. Sci. Technol. 2006, 58, 2105–2125. [Google Scholar] [CrossRef]
  61. Martín-Martín, A.; Orduña-Malea, E.; Ayllón, J.M.; Delgado-López-Cózar, E. Back to the past: On the shoulders of an academic search engine giant. Scientometrics 2016, 107, 1477–1487. [Google Scholar] [CrossRef] [Green Version]
  62. Van-Aalst, J. Using Google Scholar to estimate the impact of journal articles in education. Educ. Res. 2010, 39, 387–400. [Google Scholar] [CrossRef] [Green Version]
  63. Jacsó, P. Testing the calculation of a realistic h-index in Google Scholar, Scopus, and Web of Science for FW Lancaster. Libr. Trends 2008, 56, 784–815. [Google Scholar] [CrossRef] [Green Version]
  64. Jacsó, P. The pros and cons of computing the h-index using Google Scholar. Online Inf. Rev. 2008, 32, 437–452. [Google Scholar] [CrossRef] [Green Version]
  65. Jacsó, P. Using Google Scholar for journal impact factors and the h-index in nationwide publishing assessments in academia –siren songs and air-raid sirens. Online Inf. Rev. 2012, 36, 462–478. [Google Scholar] [CrossRef]
  66. Martín-Martín, A.; Orduña-Malea, E.; Harzing, A.W.; Delgado-López-Cózar, E. Can we use Google Scholar to identify highly-cited documents? J. Informetr. 2017, 11, 152–163. [Google Scholar] [CrossRef] [Green Version]
  67. Farhadi, H.; Salehi, H.; Yunus, M.; Aghaei-Chadegani, A.; Farhadi, M.; Fooladi, M.; Ale-Ebra-him, N. Does it matter which citation tool is used to compare the h-index of a group of highly cited researchers? Aust. J. Basic Appl. Sci. 2013, 7, 198–202. Available online: https://ssrn.com/abstract=2259614 (accessed on 23 January 2021).
  68. Marks, T.; Le, A. Increasing article findability online: The four Cs of search engine optimization. Law Libr. J. 2017, 109, 83. [Google Scholar] [CrossRef]
  69. Kearl, M.; Noteboom, C.; Tech, D. A proposed improvement to Google Scholar algorithms through broad topic search emergent research forum paper. Fac. Res. Publ. 2017, 6, 1–5. Available online: https://scholar.dsu.edu/bispapers/6 (accessed on 23 January 2021).
  70. Localseoguide. How Have Local Ranking Factors Changed? 2019. Available online: https://www.localseoguide.com/guides/local-seo-ranking-factors/ (accessed on 1 July 2020).
  71. Searchmetrics. Rebooting for Relevance. 2016. Available online: https://www.searchmetrics.com/knowledge-hub/studies/ranking-factors-2016/ (accessed on 1 July 2020).
  72. MOZ. Search Engine Ranking Factors 2015. Available online: https://moz.com/search-ranking-factors/correlations (accessed on 1 July 2019).
  73. Dave, D. 11 Things You Must Know About Google’s 200 Ranking Factors. 2018. Available online: https://www.searchenginejournal.com/google-200-ranking-factors-facts/265085/ (accessed on 10 September 2019).
  74. Chariton, R. Google Algorithm—What Are the 200 Variables? 2004. Available online: https://www.webmasterworld.com/google/4030020.htm (accessed on 10 September 2019).
  75. Wiggers, K. Google Details How It’s Using AI and Machine Learning to Improve Search. 2020. Available online: https://venturebeat.com/2020/10/15/google-details-how-its-using-ai-and-machine-learning-to-improve-search/ (accessed on 10 July 2020).
  76. Google. About Google Scholar. Available online: http://scholar.google.com/intl/en/scholar/about.html (accessed on 1 July 2019).
  77. Mayr, P.; Walter, A.-K. An exploratory study of google scholar. Online Inf. Rev. 2007, 31, 814–830. [Google Scholar] [CrossRef] [Green Version]
  78. Gielen, M.; Rosen, J. Reverse Engineering the YouTube, tubefilter.com. 2016. Available online: http://www.tubefilter.com/2016/06/23/reverse-engineering-youtube-algorithm/ (accessed on 1 July 2019).
  79. Harzing, A.-W. Publish or Perish. 2016. Available online: https://harzing.com/resources/publish-or-perish (accessed on 1 July 2019).
  80. Harzing, A.W. The Publish Or Perish Book: Your Guide to Efective and Responsible Citation Analysis; Tarma Software Research Pty Ltd.: Melbourne, Australia, 2011; pp. 39–342. Available online: https://EconPapers.repec.org/RePEc:spr:scient:v:88:y:2011:i:1:d:10.1007_s11192-011-0388-8 (accessed on 1 July 2019).
  81. R Core Team. R: A Language and Environment for Statistical Computing. Available online: https://www.R-project.org (accessed on 1 July 2019).
  82. Revelle, W. Psych: Procedures for Personality and Psychological Research, Northwestern University. 2017. Available online: https://www.scholars.northwestern.edu/en/publications/psych-procedures-for-personality-and-psychological-research (accessed on 1 July 2019).
  83. Lemon, J. Plotrix: A package in the red light district of R. R News 2006, 6, 8–12. [Google Scholar]
Figure 1. Scatter plot of searches by year.
Figure 1. Scatter plot of searches by year.
Futureinternet 13 00031 g001
Figure 2. Scatter plot of searches by author.
Figure 2. Scatter plot of searches by author.
Futureinternet 13 00031 g002
Figure 3. Scatter plot of searches by keyword.
Figure 3. Scatter plot of searches by keyword.
Futureinternet 13 00031 g003
Figure 4. Scatter plot of all searches.
Figure 4. Scatter plot of all searches.
Futureinternet 13 00031 g004
Table 1. Sample of searches carried out and the percentage of documents retrieved written in languages other than English.
Table 1. Sample of searches carried out and the percentage of documents retrieved written in languages other than English.
Percentage of Documents in Languages other than English
KeywordAuthorYear
Search% Non-EnglishSearch% Non-EnglishSearch% Non-English
capital4.0cruz7.120007.5
combustible7.1diaz6.220017.2
crisis2.7flores5.820027.4
cultural4.9garcia15.720038.6
federal3.7gomez8.020048.7
festival3.8gutierrez5.220058.6
final2.5hernandez7.820068.2
general4.6martinez13.520079.9
idea5.9morales5.220087.9
invisible3.7ortiz6.920098.1
moral4.2perez6.120108.0
popular8.6ramirez5.320118.8
social3.0ramos10.920127.2
terror4.2ranchez5.6201310.7
total2.5torres6.320149.6%
Average4.4% 7.7% 8.4%
Table 2. Number of documents ranked above 900 in the case of searches by year.
Table 2. Number of documents ranked above 900 in the case of searches by year.
Documents Ranked above 900 in Searches by Year
YearAll ResultsResults for EnglishResults for Non-English
#Items>900%#Items>900%#Items>900%
2000995959.5920232.5757296.0
2001993939.4922262.8716794.4
2002995959.5921262.8746993.2
2003993939.4908131.4858094.1
2004989899.090370.8868295.3
2005993939.4908161.8857790.6
2006994949.5913171.9817795.1
2007985858.688830.3978284.5
2008996969.6917192.1797797.5
2009989899.0909121.3807796.3
2010994949.5915202.2797493.7
2011987878.890050.6878294.3
2012991919.2920262.8716591.5
2013982828.487710.11058177.1
2014988888.989340.4958488.4
Total14,86413649.213,6142181.61250114692.1
Table 3. Number of documents ranked above 900 in the case of searches by author.
Table 3. Number of documents ranked above 900 in the case of searches by author.
Documents Ranked above 900 in Searches by Author
AuthorAll ResultsResults for EnglishResults for Non-English
#Items>900%#Items>900%#Items>900%
cruz100010010.0929323.4716895.8
diaz100010010.0938424.5625893.5
Flores100010010.0942475.0585391.4
garcia100010010.084350.61579560.5
gomez100010010.0920272.9807391.3
gutierrez100010010.0948505.3525096.2
hernandez99910010.0921283.0787292.3
martinez994959.686060.71348966.4
morales997989.8945525.5524688.5
ortiz100010010.0931343.7696695.7
perez998999.9937404.3615996.7
ramirez100010010.0947505.3535094.3
ramos99910010.089040.41099688.1
sanchez99910010.0943505.3565089.3
torres100010010.0937404.3636095.2
Total14,986149210.013,8315073.7115598585.3
Table 4. Number of documents ranked above 900 in the case of searches by keyword.
Table 4. Number of documents ranked above 900 in the case of searches by keyword.
Documents Ranked above 900 in Searches by Keyword
KeywordAll ResultsResults for EnglishResults for Non-English
#Items>900%#Items>900%#Items>900%
capital100010010.0960636.6403792.5
combustible990919.2924353.8665684.8
crisis100010010.0973757.7272592.6
cultural99910010.0950565.9494489.8
exterior100010010.0908171.9928390.2
federal100010010.0963656.7373594.6
festival100010010.0962767.9382463.2
final100010010.0975788.0252288.0
general99910010.0953596.2464189.1
idea99910010.0940434.6595796.6
invisible100010010.0963656.7373594.6
moral100010010.0958616.4423992.9
popular100010010.0914212.3867991.9
social100010010.0970717.3302996.7
terror100010010.0958646.7423685.7
Total14,98714919.914,2718495.971664289.7
Table 5. Number of documents by band of ranking in the results list.
Table 5. Number of documents by band of ranking in the results list.
Documents by Band of Ranking
Band of Ranking All ResultsResults for EnglishResults for Non-English
#Items%Items#Items%Items#Items%Items#Expected Items *%Expected Items *
from 1 to 100450010.0448010.7200.73819.0
from 101 to 200450010.0447710.7230.835111.2
from 201 to 300450010.0448210.8180.62869.2
from 301 to 400450010.0448510.8150.633810.9
from 401 to 500450010.0448110.8190.732310.4
from 501 to 600450010.0447710.7230.837612.0
from 601 to 700450010.0448010.7200.633710.8
from 701 to 800450010.0447410.7260.92849.1
from 801 to 900450010.0433110.41693.42457.9
from 901 to 10004,3379.715493.62,78891.02006.4
Total: from 1 to 100044,837100.041,716100.03,121100.03121100.0
from 1 to 209002.08932.170.21003.2
* Expected outcomes if the algorithm was applied homogeneously. Estimates based on the rank positions by number of citations of the documents in English.
Table 6. Correlations between ranking by citations and the GS native ranking for searches by year.
Table 6. Correlations between ranking by citations and the GS native ranking for searches by year.
Correlations for Searches by Year
YearAll ResultsResults for EnglishResults for Non-English
#ItemsRp-Value#ItemsRp-Value#ItemsRp-Value
20009950.88820.000009200.99510.00000750.31070.00666
20019930.88580.000009220.97730.00000710.34540.00318
20029950.86530.000009210.98560.00000740.26600.02200
20039930.85010.000009080.97920.00000850.27040.01231
20049890.82590.000009030.98440.00000860.31700.00294
20059930.85220.000009080.98640.00000850.20980.05394 *
20069940.83910.000009130.99070.00000810.38940.00033
20079850.81580.000008880.96870.00000970.37510.00015
20089960.85290.000009170.98670.00000790.33700.00239
20099890.85280.000009090.98850.00000800.38100.00049
20109940.85570.000009150.99280.00000790.32910.00306
20119870.85920.000009000.99260.00000870.25990.01506
20129910.86400.000009200.98970.00000710.23470.04885
20139820.86580.000008770.97900.000001050.44300.00000
20149880.84310.000008930.98170.00000950.33740.00083
Total14,864 13,614 1,250
Median9960.87670.000009620.93220.000001820.34980.00000
* Not statistically significant.
Table 7. Correlations between ranking by citations and the GS native ranking for searches by author.
Table 7. Correlations between ranking by citations and the GS native ranking for searches by author.
Correlations for Searches by Author
AuthorAll ResultsResults for EnglishResults for Non-English
#ItemsRp-Value#ItemsRp-Value#ItemsRp-Value
cruz10000.67290.000009290.82620.00000710.36790.00159
diaz10000.78210.000009380.89540.00000620.10470.41809 *
Flores10000.77960.000009420.91600.00000580.13900.29822 *
garcia10000.54870.000008430.83250.000001570.42650.00000
gomez10000.70420.000009200.85450.00000800.04520.69062 *
gutierrez10000.79280.000009480.89780.00000520.26540.05726 *
hernandez9990.72160.000009210.88130.00000780.30700.00625
martinez9940.62070.000008600.88850.000001340.44150.00000
morales9970.85780.000009450.92860.00000520.27540.04813
ortiz10000.77810.000009310.88730.00000690.06450.59849 *
perez9980.75370.000009370.88620.00000610.17840.16898 *
ramirez10000.80820.000009470.89730.00000530.20300.14494 *
ramos9990.69480.000008900.91400.000001090.15560.10608 *
sanchez9990.75560.000009430.86820.00000560.06800.61837 *
torres10000.76240.000009370.820.00000630.20430.10821*
Total14,986 13,831 1155
Median10000.89320.000009790.95090.000002110.10960.11232 *
* Not statistically significant.
Table 8. Correlations between ranking by citations and the GS native ranking for searches by keyword.
Table 8. Correlations between ranking by citations and the GS native ranking for searches by keyword.
Correlations for Searches by Keywords
KeywordAll Results Results for EnglishResults for Non-English
#ItemsRp-Value#ItemsRp-Value#ItemsRp-Value
capital10000.44180.000009600.49390.00000400.09080.57740 *
combustible9900.75320.000009240.85680.00000660.30470.01287
crisis10000.49500.000009730.51990.0000027−0.17130.39291 *
cultural9990.37920.000009500.41780.0000049−0.02010.89107 *
exterior10000.62860.000009080.72930.0000092−0.07050.50417 *
federal10000.58720.000009630.57360.0000037−0.04450.79361 *
festival10000.84460.000009620.86610.00000380.35350.02948
final10000.57700.000009750.58740.00000250.05140.80732 *
general9990.40120.000009530.41610.00000460.16980.25934 *
idea9990.79660.000009400.88290.00000590.28750.02725
invisible10000.69220.000009630.74920.00000370.09340.58233 *
moral10000.55020.000009580.58680.0000042−0.00860.95691 *
popular10000.61210.000009140.67630.0000086−0.00220.98379 *
social10000.23350.000009700.25610.00000300.60760.00037
terror10000.77190.000009580.78720.00000420.01890.90535 *
Total14,987 14,271 716
Median10000.91800.000009960.94120.000001510.05490.50301 *
* Not statistically significant.
Table 9. Correlations between median rank positions by citations and the GS native ranking for all searches conducted.
Table 9. Correlations between median rank positions by citations and the GS native ranking for all searches conducted.
Correlations for All Searches
All ResultsResults for EnglishResults for Non-English
#ItemsRp-Value#ItemsRp-Value#ItemsRp-Value
All searches10000.91710.00009980.97970.00003210.18030.0012
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rovira, C.; Codina, L.; Lopezosa, C. Language Bias in the Google Scholar Ranking Algorithm. Future Internet 2021, 13, 31. https://doi.org/10.3390/fi13020031

AMA Style

Rovira C, Codina L, Lopezosa C. Language Bias in the Google Scholar Ranking Algorithm. Future Internet. 2021; 13(2):31. https://doi.org/10.3390/fi13020031

Chicago/Turabian Style

Rovira, Cristòfol, Lluís Codina, and Carlos Lopezosa. 2021. "Language Bias in the Google Scholar Ranking Algorithm" Future Internet 13, no. 2: 31. https://doi.org/10.3390/fi13020031

APA Style

Rovira, C., Codina, L., & Lopezosa, C. (2021). Language Bias in the Google Scholar Ranking Algorithm. Future Internet, 13(2), 31. https://doi.org/10.3390/fi13020031

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop