Keywords

1 Introduction

When promoting their own research, researchers need to show their own research and superiority by studying related research. From an enormous number of academic papers, it is necessary for researchers to quickly and appropriately find existing studies related to their own research, that is, in a way that emphasizes the important papers. Studies of research facilities and researchers that we know as rivals are capable of checking papers, but there is always a need to further investigate related papers. Many academic paper databases provide a search function using query keywords, and it is possible to select papers according to the query. For example, Scopus provides citation numbers of papers obtained through searches as additional information for paper selection. Many papers are still selected, but since selected papers can be ranked using citation counts, many researchers check papers according to their ranking. It is necessary to check the author, title, and abstract and to obtain the main body of the paper if necessary to determine the relevance and importance of the research. The appropriate ranking of papers is an important technique that reduces waste. While there are some criticisms, the number of citations still holds an important position as a method of directly evaluating the value of a paper.

However, there is no guarantee that the number of citations is an appropriate criterion. Even if a paper has an unimportant word as a query keyword, many citations will give it a high evaluation. We need to read such a paper and remove it. Investigation of a paper will become more efficient if there is an evaluation index that is more appropriate than the number of citations.

We propose the Focused Citation Count (FCC) [9] that utilizes only appropriately cited papers in order to find other relevant research. We propose this method because we think that a paper cited from ones without appropriate content is inappropriate. Of course, such a judgment is not always easy to make. However, it is possible to make a statistical decision from many citations.

However, the FCC is easily affected by the size of the original number of citations. In the case of a paper with a large number of citations, even if the relevance is relatively small, the FCC value tends to be large. For example, we can simply show the relationship with the field you are looking for in a percentage and assume that it denotes the proportion of citations from that field. I would like to evaluate paper (A) when there is a paper (B) with a relation of 90% to paper (A) with a 10% relationship. However, if the citation number of paper (A) is 10 and the citation number of the paper (B) is 100, the FCC is 9 in paper (A) and 10 in the paper (B), and the evaluation is reversed.

In this study, we propose an enhanced method of FCC which improves this point. In order to strengthen the evaluation of relevance between the field under investigation and the thesis, we aim to reduce the influence of the number of citations by emphasizing the proportion of quotations from related fields. Experiments are carried out using actual thesis data, and the effectiveness of the proposed method is tested.

2 Focused Citation Count and Its Improvement

2.1 Focused Citation Count

FCC [9] is an evaluation index that modifies the number of citations to values useful for the proper selection of papers, by limiting the papers used for counting citations to only papers suitable for research purposes. Various methods are considered for restricting papers for this purpose, but in this study, we limit papers cited using query keywords used to search for papers. This is formulated as follows.

Let FP(q) be a set of papers selected by the query keyword q in the target where all papers are set as A. Next, let CP(p) be a set of papers citing the paper p. That is, \(PF(q)\subset A, CP(p)\subset A.\)

The total citation count cc(p) of paper p is: \(cc(p) = \left| CP(p) \right| .\)

Let CFP(pq) be the paper set selected by the query keyword q out of the paper set CP(p) citing the paper p. That is, \( CFP(p,q) = CP(p) \cap FP(q). \)

The value fcc(pq) of the FCC which is the evaluation index is obtained by Eq. (1).

$$\begin{aligned} fcc(p,q) = \left| CFP(p,q) \right| = \left| CP(p) \cap FP(q) \right| . \end{aligned}$$
(1)

2.2 Basic Idea of Improvement

Let us suppose that we are looking for related papers in a certain field. With respect to the research field of the papers citing the finished paper, I will show the proportion in which the field of both papers is the same with a simple numerical value. Suppose there is a paper (A) with a citation rate of 90% from the same field and a paper (B) with a rate of 10%. If the citation number of paper (A) is 10 and the citation number of paper (B) is 100, the FCC is 9 for paper (A) and 10 for paper (B). That is, the FCC of paper (B) is higher. However, from the viewpoint of expertise, I would like to evaluate whether paper (A) is more useful.

To make this possible, we propose using the ratio of citation counts from related fields for evaluation. In other words, since a paper with a high proportion of FCC to CC (citation count) is a paper that ought to be highly evaluated, we think that an evaluation combined with FCC and percentages could solve the problem described in this section. Proposals for concrete calculation methods are provided in the next section.

2.3 Improvement of Focused Citation Count

First, we calculate the ratio r of the number of CP(p) and CFP(pq) as the rate of the relevant field of the paper citing paper p. That is,

$$r(p,q)=\frac{\left| CFP(p,q) \right| }{\left| CP(p) \right| }=\frac{fcc(p,q)}{cc(p)}.$$

In order to combine the FCC with the r as weight, we propose a new evaluation index, Revised FCC (RFCC), by multiplying F(pq) by r(pq) raised to \(\alpha \). The value rfcc(pq) of the RFCC is given by the following equation:

$$rfcc(p,q) = fcc(p,q) \times r^\alpha (p,q).$$

However, \(\alpha \) is a parameter for adjusting the weight. In this paper, we set \(\alpha =1\).

3 Evaluation

3.1 Gathering Paper Data and Basic Analysis

In this section, we explain the collection method of the papers used for analysis, and conduct basic analyses of the data.

The data was gathered from Scopus. In this experiment, “bibliometrics” was chosen as a query keyword, and 10,186 papers published from 1976 to 2015 were gathered using search API. This data is written in JSON format. The items are as follows: “Content Type,” “Search identifier,” “Complete author list,” “Resource identifiers,” “Abstract Text,” “First author,” “Page range,” “SCOPUS Cited-by URI,” “Result URL,” “Document identifier,” “Publication date,” “Source title,” “Article title,” “Cited by count,” “ISSN,” “Issue number,” and “Volume.”

Table 1. Number of papers

Although 4,533 papers have no citation at all, there is also a paper with 4,483 citations. They are cited from 258,332 papers in total. Table 1 shows a part of the number of citations of the papers. It has the number of citations of the top 10 papers, and of the bottom 10 papers.

Fig. 1.
figure 1

Number of papers in each citation count

Figure 1 plots the data of Table 1 as a log-log graph. In this graph, the frequency of citations seems to follow the power law.

Next, I gathered papers citing each paper stated in the acquired JSON data. Information on citing papers is posted in the URL indicated in the “link” of a JSON item. Since information on the URL cannot be acquired with the API, we obtained the HTML file using the “wget” of the UNIX command. Information on 20 citing papers is published in one HTML file. However, as can be seen from Table 1, there were also citing papers exceeding 20. Regarding those, it was necessary to repeatedly obtain HTML files while changing the “wget” parameters.

Since there were 4,533 papers without a citation, we obtained the papers that cited 13,667 papers. The information on 116,743 papers on 13,667 cited papers was obtained through the execution of 10,719 wget. There were 62,265 papers when duplication was removed.

3.2 Evaluation of Revised Focused Citation Count

For the papers collected by Sect. 3.1, the following three rankings were applied. First, the top 20 papers are shown in Table 2 with the citation count indicated by Scopus. Additionally, Table 3 shows the top 20 papers in a ranking using FCC [9]. Furthermore, Table 4 shows the top 20 papers in the Revised Focused Citation Count (RFCC) proposed in this study. In these tables, “Paper ID” is the paper ID (eid) of Scopus. By replacing the \({<}\)eid\({>}\) part of the URLFootnote 1 with this ID, it is possible to acquire the data of the corresponding thesis.

Table 2. Top 20 Papers sorted by CC
Fig. 2.
figure 2

Presicion@N by CC, FCC, and RFCC

Table 3. Top 20 Papers sorted by FCC
Table 4. Top 20 Papers sorted by Revised FCC

The extraction precision of the top 20 papers extracted by CC, FCC, and RFCC was evaluated as follows. Two testers judged whether they were appropriate as papers on “bibliometrics.” They gave a rating of “1” to appropriate papers and “0” to inappropriate papers. We totaled the judgments as the number of votes, and calculated Precision@N.

The result of the CC is shown in Table 2, the result of the FCC is shown in Table 3, and the result of the RFCC is shown in Table 4. Moreover, the graph of Precision@N is shown in Fig. 2.

The result of CC was shown in Table 2, the result of FCC was shown in Table 3, and the result of RFCC was shown in Table 4. Moreover, the graph of Precision@N is shown in Fig. 2.

4 Discussion

Despite having good results in preliminary sample experiments, the RFCC could not demonstrate performance exceeding the FCC in the experiment conducted in this study. This seems to be due to the fact that the FCC shows sufficient performance, rather than the RFCC not working well. Moreover, in this experiment, using a single query keyword “bibliometrics” with a wide range of subjects may be a contributing factor.

Depending on the academic field to be searched and the kind and number of query keywords, the performance of the FCC and RFCC may differ from the results of this study. Furthermore, having two testers is a small number and may lead to biased results. As a future task, it is necessary to conduct more detailed evaluation experiments.

5 Related Work

Objective evaluation of a paper by the contents is difficult. Therefore, it is common to instead perform an evaluation using the following data of a paper: the evaluation of the journal in which the paper was published, and the paper’s citation count.

The Journal Impact Factor (JIF) [2, 3] is one of the most popular evaluation measures of scientific journals. Thomson Reuters updates and provides the scores for journals in annual Journal Citation Reports (JCR). The JIF of a journal describes the citation counts of an average paper published in the journal. JIF is considered to be the de facto standard to evaluate not only a journal, but also a researcher, research organization, and paper.

However, some problems were pointed out with JIF. Pudovkin and Garfield [8] pointed out that JIF is not appropriate to be used as a measure to compare different disciplines. Modification of JIF by normalization has been studied as one of the key issues [6, 8]. Bergstrom proposed EigenFactor [1] which solves the problem of JIF by adjusting the weight of citations. Nakatoh et al. [7] proposed to combine the relatedness of a journal to the user’s query with JIF.

The citation count is used as more direct criteria of a paper. However, some researchers have pointed out problems. Martin [5] reported that the citation count gained support as criteria. Kostoff [4] showed that the citation count as a criterion of research evaluation has the following characteristics: (a) theoretical correlation is not necessarily between a citing paper and the cited paper, (b) incorrect research may be cited, (c) a methodical paper is easy to cite, and (d) the citation count will be raised by self-citation.

Our concern is the citation count not as criteria of performance evaluation, but as criteria for selecting papers. However, the problem noted by Kostoff is common to both. The solution proposed by our study is including the quality of a citation in the evaluation. Eliminating the citation from papers with a low relation decreases the influence of the problems of (a) and (b). We think that this enables the selection of more appropriate papers.

6 Conclusion

When conducting a research survey, it is very important to quickly find suitable papers. It is common practice by researchers to select relevant papers by searching with query keywords, ranking papers by citation number, and checking in order from the highest ranked papers. However, if a paper that had a query keyword as a non-primary word has many citations, it would impede a researcher’s ability to quickly find the appropriate paper. We have already proposed a Focused Citation Count (FCC) that supports the finding of suitable papers by setting the number of citations as a more appropriate evaluation index by properly focusing on cited papers which are the sources of citation counts. In this study, we proposed an improved method of FCC. An empirical evaluation of “bibliometrics” related to 10,186 papers showed that the FCC method was effective, but that the proposed RFCC method was not as effective. As a future task, a more detailed evaluation is necessary.