1 Introduction

In the last decade and a half so-called topic modelling has emerged as a powerful statistical paradigm for the automatic semantic analysis of large collections of documents. Topic models as their name suggests can be seen as formalizations of the colloquial understanding of ‘topics’ addressed in a piece of text. More specifically, in this context a topic becomes a probability distribution over a fixed vocabulary of words (or more generally terms). Thus an example of a particular topic may be:

$$\begin{aligned} \underbrace{ \begin{array}{ccccc} 0.2 &{} 0.15 &{} 0.11 &{} 0.06 &{} \ldots \\ \textit{autistic} &{} \textit{child} &{} \textit{education} &{} \textit{needs} &{} \ldots \\ \end{array} }_\text {vocabulary size} \end{aligned}$$
(1)

where the upper row contains probabilities which correspond to the vocabulary words shown in the bottom row. Using higher-order semantic understanding a human interpreting this formal representation of a topic may describe it as capturing a discussion of educational needs of children with ASD although it should be noted that such interpretation may not always be straightforward [21]. Examples of topics extracted from real data (see Sect. 4.1.3) and visualized as so-called word clouds, with font size encoding the corresponding word probabilities, are shown in Fig. 1. Though originally developed for text analysis, topic models have since been successfully applied in a variety of other fields by suitably generalizing the concept of words [61, 65].

Fig. 1
figure 1

An illustration of typical topics discovered by the proposed method in a single epoch, shown as size-coded word clouds—a larger font indicates a proportionally more probable term (different colours are merely used for easier visualization and encode no information pertaining to the corresponding terms themselves). Topic 1 can be readily related to the persistent myth of a link between childhood vaccination and autism development as well as related issues concerning the effects of thimerosal and mercury [38], Topic 2 to the incident involving Chris Tuttle, a Wegmans employee with Asperger’s syndrome who attracted media attention in November 2013, Topic 3 to medical literature on autism, children, and brain development, and Topic 4 to the schooling of children with learning disabilities. a Example topic 1, b example topic 2, c example topic 3, d example topic 4 (colour figure online)

In more specific technical terms, topic models are probabilistic clustering approaches for grouped data where each group is a collection of data points. In text processing, each group is a document and data points are words that construct the documents. These methods find the hidden clusters in the data. Most research on topic modelling to date has focused on the analysis in the context of static corpora, that is, document collections which do not possess a temporal structure. In such collections documents are said to be interchangeable. The key techniques dominating this domain are Bayesian nonparametric inference algorithms and the latent Dirichlet allocation (LDA), in particular, first described by Blei et al. [20] and subsequently extended in a variety of ways. Indeed at the bottom-most level the present paper uses a model based on the hierarchical Dirichlet process (HDP) which is one of the aforementioned extensions. Both LDA and HDP are explained in some detail in Sect. 3.1.

However, in many problems of practical significance, it is not only the instantaneous topic structure that is of interest—the change in this structure over time too often conveys important information and insight. For this reason in recent years the problem of temporal topic modelling has been attracting an increasing amount of research attention [2, 4, 12, 13, 16, 25]. Indeed the focus of the present paper is on temporally changing corpora. At the heart of the method that we describe is an automatically constructed and temporally constrained graph superimposed on topics extracted from short-time epochs within which the corpus can be considered as being static.

In particular, the work described herein extends our contribution first described in [12]. Retaining the same structural framework, in the present paper we analyse in detail the effects of different pruning parameters in the construction of the graph superimposed over the extracted topics. Specifically we consider different choices of inter-topic distance measures and the process of selecting an appropriate pruning threshold given a specific distance measure. In addition, we describe extended experiments on the collection of abstracts of scholarly articles on the autism spectrum disorder (ASD), gathered by ourselves and made publicly available, as well as additional experiments on a newly gathered collection of abstracts of scholarly articles in the highly active sphere of work on the metabolic syndrome. This new collection of documents will also be made publicly available.

The remainder of the present paper is structured as follows. In the next section we review relevant previous work, both on topic modelling in general as well as on temporal modelling which is the focus of our work. Then in Sect. 3 we first describe the key prerequisite modelling techniques which our contributions build upon (Sect. 3.1), followed by the detail on the proposed modelling framework (Sect. 3.2). Section 3.2 introduces our main technical contributions. Specifically these are our general temporal model based on what we term the temporal similarity graph, and the two key aspects involved in the construction of this graph: the choice of an inter-topic similarity measure (Sect. 3.2.1) and a graph pruning process (Sect. 3.2.2). The proposed method is extensively evaluated in Sect. 4. Section 4.1 motivates our focus on biomedical documents used in the evaluation with Sect. 4.1.1 summarizing previous work in the specific domain of text mining, Sect. 4.1.2 provides a brief synopsis of the medical conditions the two specific document data sets we used for evaluation relate to, and Sect. 4.1.3 describes the data sets themselves, the manner in which data were collected, as well as the pre-processing thereof. Section 4.2 details different inter-topic similarity measures we compared. Our experiments and the associated findings and discussion are presented in Sect. 4.3, with Sect. 4.3.1 focusing on quantitative elements of the evaluation and Sect. 4.3.2 on qualitative analysis. Finally, the contributions and the key findings of the present work are summarized in Sect. 5.

2 Previous work

Owing to its growing popularity in recent years, the existing literature on probabilistic topic modelling is already vast. Hence a comprehensive survey of the field is out of the scope of the present paper. Herein we overview the broad and most influential research directions, with a particular focus on techniques of direct relevance to the methodology described in the proposed paper. In particular we direct our attention first to latent topic models which have dominated the field in the last decade, and then on biomedical text mining, given the application domain within which our framework is evaluated in Sect. 4.

Topic models in modern machine learning are often described as latent probabilistic models. The attribute ‘latent’ is intended to capture the nature of inferred topics—these are hidden variables in the sense that they are not explicit in the observable data itself. Arguably the sought after ‘true’ topic structure is also neither objective nor accessible even in principle. On the other hand, the attribute ‘probabilistic’ conveys the inherent aspect of the aforementioned models whereby modelling imperfections as well as ambiguities in data are handled through the use of probability distributions which readily accommodate uncertainty. Therefore we start our discussion with the simplest latent topic models which formalize the key ideas in the field and underlie much of the subsequent, more complex models.

2.1 Latent topic models

An important early topic modelling approach comes in the form of so-called latent semantic indexing (LSI) [24] which remains popular. Two notable limitations of LSI are its inability to deal effectively with polysemy and to produce an explicit description of the latent space. A probabilistic improvement overcomes these by explicitly characterizing the latent space with semantic topics, and by employing a probabilistic generative model that addresses the polysemy problem [36]. Nevertheless, probabilistic LSI is prone to parameter overfitting caused by an uncontrolled growth in the number of parameters as the document corpus is increased. In addition, the necessary assignment of probabilities to documents is a non-trivial task [20].

The recently proposed latent Dirichlet allocation (LDA) method [20] overcomes the overfitting problem by adopting a Bayesian framework and a generative process at the document level. While LDA has quickly become a standard tool for topic modelling, it too experiences challenges when applied on real-world data. In particular, being a parametric model the number of desired output topics has to be specified in advance. The HDP model as the nonparametric counterpart of LDA was introduced by Teh et al. [59] and addressed this limitation by using a hierarchical Dirichlet process (as opposed to a Dirichlet distribution) as the prior on topics. Therefore, each document is modelled using an infinite mixture model, allowing the data to inform the complexity of the model and infer the number of resulting topics automatically. We discuss this model in further detail in Sect. 3.

2.1.1 Temporal topic modelling

A notable limitation of most models described in the existing literature lies in their assumption that the data corpus is static; this includes those based on LDA mentioned previously, or the hierarchical Dirichlet process described in detail in the next section. Here the term ‘static’ is used to describe the lack of any associated temporal (or indeed sequential) information associated with the documents in a corpus—the documents are said to be exchangeable [18]. However, in many practical applications documents are added to the corpus in a temporal manner and their ordering has significance, i.e. the documents are non-exchangeable. As a consequence, the topical structure of the corpus changes over time. For example, the corpus of scholarly literature on a particular subject is a growing corpus which by its very nature exhibits significant changes in its topic structure over time [6, 26, 27]: new ideas emerge, old ideas are refined, novel discoveries result in multiple ideas being related to one another thereby forming more complex concepts or a single idea multifurcating into different ‘sub-ideas’ which are thereafter investigated with some degree of independence, etc. A good example from a different realm can be readily found in the corpus of social media contributions, such as Twitter [13]. With an even faster pace here too complex topic changes can be observed, with novel topics of conversation being instigated by, e.g. ‘real-world’ events (such as epidemics, terrorist attacks, or developments in the world of popular culture), changed by the contributions of other users, split into new topics, merged with others.

The assumption made by all previous work, and indeed adopted by us, is that documents are not exchangeable at large temporal scales but are at short-time scales, thus treating the corpus as temporally locally static. The scale at which this assumption can be considered as valid is clearly application and corpus dependent and is an important consideration. Indeed the present paper investigates this in detail.

The existing work on temporal topic modelling can be divided into two groups of approaches both of which can be based on parametric [18, 63, 64] or nonparametric [48, 68] techniques, the former suffering from the limitation that they contain free parameters which must be set a priori. Methods of the first group discretize time into epochs, apply a static topic model to each epoch, and by making the Markovian assumption relate the parameters of each epoch’s topic model to those of the epochs adjacent to it in time [18, 48, 63, 68]. While the approach we propose in this paper adopts the idea of time discretization, it diverges in its other features from this group of methods thereafter. In particular, instead of employing the Markovian assumption we describe a novel structure in form of a temporal similarity graph, which gives our method greater flexibility, as described in detail in the next section. The second group of methods in the literature regard document timestamps as observations of a continuous random variable [25, 64]. This assumption severely limits the type of topic changes which can be described. For example, as opposed to our model, these models are not capable of describing the evolution of topics, or their splitting and merging, and are rather constrained to tracking simple topic popularity.

3 Proposed framework

In this section we present the technical contribution of the present work. We begin by reviewing the relevant theory underlying Bayesian mixture modelling, LDA, and HDP that plays a central role in the proposed framework. Then we turn our attention to the novel contribution of our work and explain how the aforementioned Bayesian techniques are extended to deal with temporally varying document corpora.

3.1 Bayesian mixture models

In recent years mixture models have become popular choices for the modelling of so-called heterogeneous data. In this context heterogeneity is taken to mean that observable data are generated by more than one process (source). One of the key challenges in the analysis of heterogenous data lies in the lack of observability of the correspondence between specific data points and their sources, i.e. it is not known which data source generated which data. Usually it is also the case that the number of sources is not known either [50]. Mixture models, and in particular mixture models enveloped within a Bayesian framework, have distinct advantages over alternative approaches as we shall explain shortly.

3.1.1 Finite mixture modelling

Finite mixture models rely on the assumption that the observed data are generated by K clusters, each cluster being associated with the parameter \(\phi _{k}\) and underlain by the probability density function \(f\left( .|\phi _{k}\right) \). An observation x is assumed to be generated by first choosing a cluster k with probability \(\pi _{k}\) followed by a random draw from the corresponding distribution described by \(\phi _{k}\). Therefore the process can be summarized by the following:

$$\begin{aligned} p\left( x|\pi _{1:K},\phi _{1:K}\right) =\sum _{k=1}^{K}\pi _{k}f\left( x|\phi _{k}\right) . \end{aligned}$$
(2)

In a Bayesian setting the model parameters (i.e. mixing proportions \(\pi _{1:K}\) and component parameters \(\phi _{1:k}\)) are further endowed by priors. Typically the symmetric Dirichlet distribution is placed on top of \(\pi _{1:K}\) and a prior on \(\phi _{1:K}\) conjugate with \(f\left( .|\phi _{k}\right) \) chosen for computational convenience.

3.1.2 Latent Dirichlet allocation

In the previous section we described how to model a group of data points with a Bayesian finite mixture model. Latent Dirichlet allocation adds a level of hierarchy on the mixing proportions to allow for the modelling of data points in groups that share a set of components.

Following the consensus in the literature we adopt the terminology used in the analysis of textual data (which is the context in which LDA was originally proposed [20]) and hereafter interexchangably refer to data points as words, their groups as documents, and mixture components as topics. The technical term ‘topic’ can be interpreted as formalizing and abstracting the colloquial notion of a topic which is understood at a higher semantic level. Therefore the modelling framework of LDA can be described by the following generative process:

$$\begin{aligned} \phi _{1:K}\sim & {} H, \end{aligned}$$
(3)
$$\begin{aligned} \pi _j\sim & {} \text {Dirichlet}\left( \alpha \right) , \end{aligned}$$
(4)
$$\begin{aligned} z_{ji}|\pi _j\sim & {} \pi _j, \end{aligned}$$
(5)
$$\begin{aligned} x_{ji}|z_{ji},\phi _{1:K}\sim & {} F\left( \phi _{z_{ji}}\right) , \end{aligned}$$
(6)

where H is the base distribution of topics, \(\alpha \) the hyperparameter of the prior on the distribution of topics within a document, \(\pi _j\) the distribution of topics in document j, and \(z_{ji}\) the topic corresponding to the ith word in the jth document. The corresponding model likelihood is:

$$\begin{aligned} p\left( w_{ji}|\alpha \right) =\int _{\pi _{j}}\int _{\phi _{1:K}}\sum _{k=1}^{K}\pi _{jk}f\left( x|\phi _{k}\right) \hbox {d}P\left( \pi _{j}\right) \hbox {d}P\left( \phi _{1:K}\right) , \end{aligned}$$
(7)

Approximation techniques such as MCMC [32] and variation Bayes [20] methods can be used for posterior inference.

3.1.3 Infinite mixture modelling

As mentioned earlier, LDA requires the number of topics to be fixed in advanced which is a serious limitation in practice. Choosing the number of topics is usually performed by examining how well the model fits a set of held-out documents. However, if a previously unseen topic has contributed in generating the held-out data, LDA is not able to infer correct parameters of that topic.

Bayesian nonparametric (BNP) methods place priors on the infinite-dimensional space of probability distributions and provide an elegant solution to this problem. Dirichlet process (DP) [28] as the nonparametric counterpart of the Dirichlet distribution and the building block of BNP allows for the model to accommodate a potentially infinite number of mixture components. The generative likelihood for a data point x in infinite mixture model is:

$$\begin{aligned} p\left( x|\pi _{1:\infty },\phi _{1:\infty }\right) =\sum _{k=1}^{\infty }\pi _{k}f\left( x|\phi _{k}\right) . \end{aligned}$$
(8)

\(\text {DP}\left( \gamma ,H\right) \) is defined as a distribution of a random probability measure G over a measurable space \(\left( \Theta ,\mathcal {B}\right) \), such that for any finite measurable partition \(\left( A_{1},A_{2},\ldots ,A_{r}\right) \) of \(\Theta \) the random vector \(\left( G\left( A_{1}\right) ,\ldots ,G\left( A_{r}\right) \right) \) is a Dirichlet distribution with parameters \(\left( \gamma H\left( A_{1}\right) ,\ldots ,\gamma H\left( A_{r}\right) \right) \). A DP generates imperfect atomic copies of its base measure H with a variance governed by its concentration parameter \(\gamma \). An alternative view of the DP emerges from the so-called stick-breaking process which adopts a constructive approach using a sequence of discrete draws [53]. Specifically, if \(G\sim \text {DP}\left( \gamma ,H\right) \) then \(G=\sum _{k=1}^{\infty }\beta _{k}\delta _{\phi _{k}}\) where \(\phi _{k}\overset{iid}{\sim }H\) Footnote 1 and \(\varvec{\beta }=\left( \beta _{k}\right) _{k=1}^{\infty }\) is the vector of weights obtained by a stick-breaking process that is \(\beta _{k}=v_{k}\prod _{l=1}^{k-1}\left( 1-v_{l}\right) \) and \(v_{l}\overset{iid}{\sim }{\text {Beta}}\left( 1,\gamma \right) \).

Owing to the discrete nature and infinite dimensionality of its draws, the DP is a highly useful prior for Bayesian mixture models. By associating different mixture components with atoms \(\phi _{k}\) of the stick-breaking process, and assuming \(x_{i}|\phi _{k}\overset{iid}{\sim }f\left( x_{i}|\phi _{k}\right) \) where \(f\left( .\right) \) is the likelihood kernel of the mixing components, we can formulate the infinite Bayesian mixture model or Dirichlet process mixture model (DPM). Approximate methods are used for posterior inference [47].

3.1.4 Hierarchical Dirichlet process mixture models

While DPM is suitable for nonparametric clustering of exchangeable data in a single group, many real-world problems are more appropriately modelled as comprising multiple groups of exchangeable data. In such cases it is usually desirable to model the observations of different groups jointly, allowing them to share their generative clusters. This idea is known as ‘sharing the statistical strength’ and it is naturally obtained by hierarchical architecture in Bayesian modelling.

Consider a collection of documents. DPM models each group with an infinite number of topics. However, it is desired for multiple group-level DPMs to share their clusters. Amongst different ways of linking group-level DPMs, HDP [57] offers an interesting solution whereby base measures of group-level DPs are drawn from a corpus-level DP. In this way the atoms of the corpus-level DP (i.e. topics in our case) are shared across the documents. Formally, if \(\mathbf {x}=\left\{ \mathbf {x}_{1},\ldots ,\mathbf {x}_{J}\right\} \) is a document collection where \(\mathbf {x}_{j}=\left\{ x_{j1},\ldots ,x_{jN_{j}}\right\} \) is the jth document comprising \(N_{j}\) words:

$$\begin{aligned} G_{0}|\gamma ,H\sim & {} \text {DP}\left( \gamma ,H\right) \end{aligned}$$
(9)
$$\begin{aligned} G_{j}|\alpha _{0},G_{0}\sim & {} \text {DP}\left( \alpha _{0},G_{0}\right) \end{aligned}$$
(10)
$$\begin{aligned} \theta _{ji}|G_{j}\sim & {} G_{j} \end{aligned}$$
(11)
$$\begin{aligned} x_{ji}|\theta _{ji}\sim & {} F\left( .|\theta _{ji}\right) \end{aligned}$$
(12)

This is illustrated schematically in Fig. 2a. Since \(G_{j}\) is drawn from a DP with base measure \(G_{0}\), it takes the same support as \(G_{0}\). Also the parameters of the group-level mixture components, \(\theta _{ji}\), share their values with the corpus-level DP support on \(\left\{ \phi _{1},\phi _{2},\ldots \right\} \). Therefore \(G_{j}\) can be equivalently expressed using the stick-breaking process as \(G_{j}=\sum _{k=1}^{\infty }\pi _{jk}\delta _{\phi _{k}}\) where \(\varvec{\pi }_{j}|\alpha _{0},\gamma \sim \text {DP}\left( \alpha _{0},\gamma \right) \) [58]. The posterior for \(\theta _{ji}\) has been shown to follow a Chinese restaurant franchise process which can be used to develop inference algorithms based on Gibbs sampling [59].

Fig. 2
figure 2

a Graphical model representation of HDP. Each box represents one document whose observed data (words) is shown shaded. Unshaded nodes represent latent variables. An observed datum \(x_{ji}\) is assigned to a latent mixture component parameterized by \(\theta _{ji}\). \(\gamma \) and \(\alpha \) are the concentration parameters and H is the corpus-level base measure. b Graphical model representation of the proposed framework. The corpus is temporally divided into \(t_{n}\) epochs and each epoch modelled using an HDP (outer boxes). Different epochs’ HDPs share their corpus-level DP and hyperparameters

3.2 Modelling topic evolution over time

Hitherto, the discussion in this section focused on the modelling of static document corpora. We now show how our work builds on top of these ideas in the existing literature and in particular how the described HDP-based model can be applied to the analysis of temporal topic changes in a longitudinal data corpus.

Like some of the previous work in this area we begin by discretizing time and dividing the literature corpus by time into epochs. Each epoch spans a certain contiguous time period and has associated with it all documents with timestamps within this period. The duration of epochs should be sufficiently short to allow the corresponding document subset to be treated as a static collection; we shall discuss this issue in more detail in the next section.

Each epoch is then modelled separately using an HDP, with models corresponding to different epochs sharing their hyperparameters and the corpus-level base measure. Hence if n is the number of epochs, we obtain n sets of topics \(\varvec{\phi }=\left\{ \varvec{\phi }_{t_{1}},\ldots ,\varvec{\phi }_{t_{n}}\right\} \) where \(\varvec{\phi }_{t}=\left\{ \phi _{1,t},\ldots ,\phi _{K_{t},t}\right\} \) is the set of topics that describe epoch t, and \(K_{t}\) their number (which is inferred automatically, as described previously). This is illustrated in Fig. 2b. In the next section we describe how given an inter-topic similarity measure the evolution of different topics across epochs can be tracked.

3.2.1 Measuring topic similarity

Our goal now is to track changes in the topical structure of a data corpus over time. The simplest changes of interest include the emergence of new topics, and the disappearance of others. More subtly, we are also interested in how a specific topic changes, that is, how it evolves over time in terms of the contributions of different words it comprises. Lastly, our aim is to be able to extract and model complex structural changes of the underlying topic content which result from the interaction of older topics. Specifically, topics, which can be thought of as collections of memes, can merge to form new topics or indeed split into more nuanced memetic collections. This information can provide valuable insight into the refinement of ideas and findings in the scientific community, effected by new research and accumulating evidence.

The key idea behind our approach stems from the observation that while topics may change significantly over time, providing that the duration of epochs is chosen appropriately, the change between successive epochs is limited. Therefore we infer the continuity of a topic in one epoch by relating it to all topics in the immediately subsequent epoch which are sufficiently similar to it under a suitable similarity measure. This can be seen to lead naturally to a similarity graph representation whose nodes correspond to topics and whose edges link those topics in two epochs which are related. Formally, the weight of the directed edge that links \(\phi _{j,t}\), the jth topic in epoch t, and \(\phi _{k,t+1}\) is set equal to \(\rho \left( \phi _{j,t},\phi _{k,t+1}\right) \) where \(\rho \) is a similarity measure. Given that in our HDP-based model each topic is represented by a probability distribution, similarity measures, such as the Bhattacharyya coefficient [5], the Jensen–Shannon divergence, and the Hellinger distance for example, are all readily adapted for use in the proposed framework as we shall demonstrate in the next section.

A conceptual illustration of a small section of a similarity graph is shown in Fig. 3. It shows three consecutive time epochs \(t-1\), t, and \(t+1\) and a selection of topics in these epochs. Graph edge weight (i.e. inter-topic similarity) is encoded by the thickness of the line representing the edge—the thicker the line, the more similar the corresponding topics are. In constructing a similarity graph we use a threshold to eliminate automatically weak edges, retaining only the connections between sufficiently similar topics in adjacent epochs. This readily allows us to detect the disappearance of a particular topic, the emergence of new topics, as well as the splitting or merging of different topics as follows:

  • New topic emergence:  If a node does not have any edges incident to it, the corresponding topic is taken as having emerged in the associated epoch; for example, in Fig. 3 the topic \(\phi _{j+2}\) can be seen to emerge during the epoch with the timestamp t.

  • Topic disappearance: If no edges originate from a node, the corresponding topic is taken to vanish in the associated epoch; for example, in Fig. 3 the topic \(\phi _{j}\) can be seen to disappear during the epoch with the timestamp t.

  • Topic evolution: When exactly one edge originates from a node in one epoch and it is the only edge incident to a node in the following epoch, the topic is understood as having evolved in the sense that its memetic content (captured by the probability distribution over the underlying vocabulary) may have changed; for example, in Fig. 3 the topic \(\phi _{j+2}\) evolves into the topic \(\phi _{k+1}\) between the epochs with the timestamps t and \(t+1\).

  • Topic splitting: If more than a single edge originates from a node, we interpret this as the corresponding topic splitting into multiple topics between the corresponding epochs and the successive epoch; for example, in Fig. 3 the topic \(\phi _{i}\) splits into topics \(\phi _{j}\) and \(\phi _{j+1}\) between the epochs with the timestamps \(t-1\) and t.

  • Topic merging: If more than a single edge is incident to a node, the topics of the nodes from which the edges originate are understood to have interacted by merging to form a new topic; for example, in Fig. 3 between the epochs with the timestamps \(t-1\) and t the topics \(\phi _{i}\) and \(\phi _{i+1}\) merge to form \(\phi _{j+1}\).

Fig. 3
figure 3

Conceptual illustration of the proposed similarity graph that models topic dynamics over time. A node corresponds to a topic in a specific epoch; edge weights are equal to the corresponding topic similarities

Lastly, observe that just as our understanding of topics at a higher semantic level would allow, the proposed framework readily models complex structural changes which involve a topic concurrently undergoing merging and splitting. For example, the topic labelled \(\phi _{j+1}\) in Fig. 3 is created though the merging of topic \(\phi _{i+1}\) and a split offshoot of the topic \(\phi _i\).

3.2.2 Automatic temporal similarity graph construction

In our previous work [12, 13] the temporal similarity graph is built in two stages. In the first stage the graph is connected fully in the sense that all pairs of topics in successive epochs are connected by edges. Then, the graph is pruned using a similarity threshold \(t_s\). In other words any edge corresponding to an inter-topic similarity lesser than \(t_s\) is removed from the graph.

A major limitation of the described pruning step is that the similarity threshold \(t_s\) is not readily interpretable and the framework provides little insight as to how the threshold should be chosen. In addition, considering that the threshold value inherently depends on the similarity measure which is used, it is not clear how two inter-topic similarity measures may be compared, i.e. how to control for the threshold in the presence of a changing similarity metric which underlies it. Hence in the present paper we describe an alternative strategy which employs a more meaningful and more interpretable manner of pruning. Firstly we consider all inter-topic similarities present in the initial fully connected graph and extract the empirical estimate of the corresponding cumulative density function (CDF). Examples of CDFs obtained using three similarity metrics on a typical epoch in our data (please see the next section for empirical analysis) are shown for illustration in Fig. 4. The very different functional forms of the three CDFs shown reflect our previous observation that pruning on the basis of a similarity threshold is inherently dependent on the employed similarity measure, which complicates any comparative analysis. Instead of using a similarity threshold herein we prune the graph based on the operating point on the relevant CDF. In other words if \(F_\rho \) is the CDF corresponding to a specific initial, fully connected graph formed using a particular similarity measure, and \(\zeta \in [0, 1]\) the CDF operating point, we prune the edge between topics \(\phi _{j,t}\) and \(\phi _{k,t+1}\) iff \(\rho (\phi _{j,t},\phi _{k,t+1}) < F^{-1}_\rho (\zeta )\). The proposed framework is succinctly summarized in Algorithm 1.

Fig. 4
figure 4

Empirical estimates of the cumulative density function (CDF) of inter-topic similarities corresponding to fully connected temporal similarity graph constructed over the same set of topics but using three different similarity measures (please see Sect. 4.2). Such CDFs are used in the proposed construction of the final temporal similarity graph, that is, in the pruning process used to derive it from the precedent, fully connected graph

figure e

4 Experimental evaluation

Having introduced the main technical contribution of our work we now analyse the performance of the proposed framework empirically on two large real-world data sets (Fig. 5).

Fig. 5
figure 5

Examples of temporal similarity graphs. All three graphs correspond to the same data corpus and are formed over the same set of topics using a similarity measure based on the Hellinger distance (please see Sect. 4.2) but using different CDF operating points \(\zeta =0.9, 0.95, 0.99\). The horizontal axis denotes time with each column of topics (dots) corresponding to a specific epoch. The size of dots represents topics codes for their popularity in the corresponding epoch. As expected, increasing \(\zeta \) results in an increase in sparsity of the final temporal graph. a \(\zeta =0.9\), b \(\zeta =0.95\), c \(\zeta =0.99\)

4.1 Evaluation data

In Sect. 2 we highlighted the corpus of published peer-reviewed scientific literature as a particularly attractive data source for the application of temporal topic models. This corpus exhibits rapid growth at an accelerating pace, and it is by its very nature characterized by high fluency of ideas. These ideas can be represented well by topics in the sense in which this term is used in the present article. Hence for the evaluation of the method proposed in the preceding section we selected two highly active research areas in the realm of biomedicine.

For the sake of completeness we next briefly survey the most influential work on the analysis of biomedical texts, then outline the key reasons for our choice of the specific research areas we focus on, and then proceed with a description of the adopted experimental methodology, a presentation of the most significant results emerging from our analysis, and the associated discussion.

4.1.1 Biomedical text mining

Most previous work on text based knowledge discovery in biomedicine to date has focused on (i) the tagging of names of entities such as genes, proteins, and diseases [54], (ii) the discovery of relationships between different entities, e.g. functional associations between genes [49], or (iii) the extraction of information pertaining to events such as gene expression or protein binding [55].

The idea that the medical literature could be mined for new knowledge is typically attributed to Swanson [56]. For example by manually examining medical literature databases he hypothesized that dietary fish oil could be beneficial for Raynaud’s syndrome patients, which was later confirmed by experimental evidence. Work that followed sought to develop statistical methods which would make this process automatic. Most approaches adopted the use of term frequencies and co-occurrences using dictionaries such as Medical Subject Headings (MeSH) [51].

Most existing work on biomedical knowledge discovery is based on what may be described as traditional data mining techniques (neural networks, support vector machines, etc); comprehensive surveys can be found in [40, 55]. The application of state-of-the-art Bayesian methods in this domain is scarce. Amongst the notable exceptions is the work by Blei et al. who showed how latent Dirichlet allocation (LDA) can be used to facilitate the process of hypothesis generation in the context of genetics [19]. Arnold et al. used a similar approach to demonstrate that abstract topic space representation is effective in patient specific case retrieval [8]. In their later work they introduced a temporal model which learns topic trends and showed that the inferred topics and their temporal patterns correlate with valid clinical events and their sequences [9]. Wu et al. used LDA for gene–drug relationship ranking [67].

4.1.2 Autism spectrum disorder and the metabolic syndrome

In this paper we evaluate our topic discovery framework on two corpora of scholarly papers. The first corpus is of abstracts of papers concerning the autism spectrum disorder, and the other of abstracts of papers related to research on the metabolic syndrome. These specific research areas are chosen for several reasons. Firstly, they concern medical issues of major practical importance—they affect (directly or indirectly) a large number of people and impose a significant financial cost both to the society as a whole and to those affected. Secondly, the understanding of mechanisms underlying both conditions have proven to pose a significant intellectual challenge. Consequently, the dominant ideas regarding the underlying causative mechanisms, modes of treatment and their efficacy, etc., are continually changing, experiencing both refinement as well as more abrupt paradigm shifts. These aspects make the chosen areas of research highly suitable for the evaluation of the framework described in the present work.

4.1.2.1 The autism spectrum disorder

Autism spectrum disorder is a lifelong neurodevelopmental disorder with poorly understood causes on the one hand, and a wide range of potential treatments supported by little evidence on the other. The disorder is characterized by severe impairments in social interaction, communication, and in some cases cognitive abilities, and typically begins in infancy or at the very latest by the age of three. ASD is recognized as comprising an aetiologically and clinically heterogeneous group of conditions whose diagnosis remains to be based solely on the complex behavioural phenotype [45]. According to the definition in the latest version (5th edition) of the Diagnostic and Statistical Manual of Mental Disorders, the autism spectrum disorder includes disorders which were previously diagnosed with more specificity as autism, Asperger syndrome, Rett syndrome, childhood disintegrative disorder, and ‘pervasive developmental disorder not otherwise specified’ [1]. Current evidence suggests that approximately 0.5–0.6% of the population is afflicted by ASD though the actual diagnosis rate is on the increase due to the broadening diagnostic criteria [10]. The condition is usually detected in early childhood when an abnormal lack of social reciprocity is observed.

Although the last few decades have seen significant progress in the study of ASD, the still relatively poorly understood aetiology of the condition, its phenotypical heterogeneity [42], and stigma associated with mental conditions [31] have all contributed to the penetration of beliefs, and behavioural and educational interventions which are often questionable [66] and poorly supported by evidence (e.g. gluten-free and casein-free diets, and cognitive behavioural therapy [23]), and sometimes outright in conflict with science [60]. For example, a recent review of early intensive behavioural and developmental interventions for young children with ASD found 1 existing study as being of good quality, 10 as fair quality, and 23 as poor quality [66]. From the public policy point of view, understanding the practices and beliefs of parents and carers of ASD-affected individuals is crucial, yet often lacking [34].

4.1.2.2 Metabolic syndrome

Much like ASD, metabolic syndrome (also known as insulin resistance syndrome and syndrome X) does not describe a single disorder but rather a cluster of interconnected health risk factors [33]. Specifically, the diagnostic criterion is the presence of at least three of the following: visceral obesity, arterial hypertension, hyperglycaemia, hypertriglyceridemia, and hypoalphalipoproteinemia [33]. MetS is recognized as a major and escalating public health challenge, chiefly in the developed world, and is thought to be caused in part by excess energy intake, and decreased energy output due to an increasingly sedentary lifestyle [41]. Metabolic syndrome is associated with an increased risk of numerous diseases and particularly notably with the development of cardiovascular disease and type 2 diabetes mellitus [3]. Approximately one-third of adults in the USA suffer from MetS [30], with the prevalence of the syndrome increasing with age [30].

4.1.3 Data collection

To the best of our knowledge there are no publicly available corpora of ASD or MetS related medical literature. Hence we collected them ourselves. These are now publicly available and can be downloaded from: https://adham.github.io/KAIS16/. The data sets, their collection, and preparation are described next.

4.1.4 Raw data collection

We used the PubMed interface to access the US National Library of Medicine and retrieve abstracts and references of life science and biomedical scholarly articles. We assumed that a paper is related to ASD or MetS, respectively, if the terms ‘autism’ or ‘metabolic syndrome’ are present in its title or abstract, and collected only papers written in English. The earliest publications fitting our criteria are by Kanner [39] on ASD and by Berardinelli et al. [11] on MetS. We collected all matching publications up to the final one indexed by PubMed on 10th May 2015, yielding a corpus of 22,508 on ASD and 31,706 publications on MetS. We used the abstract text to evaluate our method.

4.1.5 Data pre-processing

Data collected in the manner described in the previous section comprises abstracts as freeform text. To prepare it for the type of analysis described in Sect. 3 we perform a series of ‘pre-processing’ steps. The goal is to remove words which are largely uninformative in any context, reduce dispersal of semantically equivalent terms, and thereafter select terms which are included in the vocabulary over which topics are learnt [14, 15].

We firstly applied soft lemmatization using the WordNet\(^\circledR \) lexicon [46] to normalize for word inflections. No stemming was performed to avoid semantic distortion often effected by heuristic rules used by stemming algorithms. After lemmatization and the removal of so-called stop words, we obtained approximately 2.2 and 3.8 million terms in the entire corpus when repetitions are counted, and 37,626 and 46,114 unique terms, for the ASD and MetS corpora, respectively. Constructing the vocabulary for our method by selecting the most frequent terms which explain 90% of the energy in a specific corpus resulted in ASD and MetS vocabularies containing 3417 and 2839 terms, respectively.

4.2 Inter-topic similarity measures

Recall that in this work topics are probability distributions over a fixed vocabulary of terms. Thus the inter-topic similarity measure used to construct our temporal graph is thus similarity measures between probability distributions. Considering that the vocabulary of terms is fixed we represent each topic, say p, using a fixed length vector:

$$\begin{aligned} p = \left[ p_1, p_2, \ldots , p_{n_v} \right] ^\mathrm{T} \end{aligned}$$
(13)

where \(n_v\) is the number of terms in the vocabulary.

For the experiments in this paper we adopted three well-known measures for quantifying the similarity (or equivalently, dissimilarity) between probability distributions representing extracted topics. The first of these is the well-known Hellinger distance [35]. For two discrete probability distributions, e.g. p and q representing two topics, it is defined as follows:

$$\begin{aligned} H(p,q) = \frac{1}{\sqrt{2}} \sqrt{ \sum _{i=1}^{n_v} (\sqrt{p_i} - \sqrt{q_i})^2 }. \end{aligned}$$
(14)

It can be readily seen that H() is symmetric and that it takes on a value between 0 and 1, with 0 signifying the greatest degree of similarity between p and q (in this case \(p=q\)) and 1 the least (in this case \(p_i> 0 \implies q_i=0 ~~\wedge ~~ q_i > 0 \implies p_i=0\)).

The second similarity measure we evaluate is the Bhattacharyya coefficient defined as [17]:

$$\begin{aligned} B(p,q) = \sum _{i=1}^{n_v} \sqrt{p_i q_i}. \end{aligned}$$
(15)

As the Hellinger distance, the Bhattacharyya coefficient is symmetric and takes on a value between 0 and 1. However note that in this case it is the maximum value of 1 which is attained when there is the greatest degree of similarity between p and q (i.e. \(p=q\)) and 0 the least (as before in this case \(p_i> 0 \implies q_i=0 ~~\wedge ~~ q_i > 0 \implies p_i=0\)). It is straightforward to demonstrate that the Bhattacharyya coefficient is related to the Hellinger distance as follows:

$$\begin{aligned} H(p,q) = \sqrt{ 1 - B(p,q) }. \end{aligned}$$
(16)

Lastly, due to its widespread use we also compare the performance of the proposed algorithm using the similarity measure often erroneously referred to as Tanimoto similarity [43] (or the Jaccard similarity [29]), defined as:

$$\begin{aligned} T(p,q) = \frac{ \sum _{i=1}^{n_v} p_i q_i }{\sum _{i=1}^{n}{p_i}^2+\sum _{i=1}^{n_v} {q_i}^2-\sum _{i=1}^{n_v}p_i q_i}. \end{aligned}$$
(17)

As the Bhattacharyya coefficient, this measure is symmetric and takes on a value between 0 and 1, with the maximum value of 1 being attained when p and q are equal, and 0 when there is no overlap between them (i.e. \(p_i> 0 \implies q_i=0 ~~\wedge ~~ q_i > 0 \implies p_i=0\)). In an effort to avoid perpetuating the aforementioned misnomer on the one hand while retaining some nomenclatural continuity with the existing literature, we will refer to this measure as quasi-Jaccard similarity.

4.3 Experiments and results

In this section we conduct experiments on the two corpora of scholarly literature described in Sect. 4.1.3, and report and discuss our results both using quantitative findings and representative qualitative examples.

4.3.1 Quantitative comparison

We started our evaluation with an experiment examining quantitative differences effected by changing different flexible parameters of the proposed method. In particular, our aim was to see how the evolving topic structure extracted by our algorithm is affected by different choices of inter-topic similarity measures (introduced and described in Sect. 3.2.1 and used to construct the temporal similarity graph as described in Sect. 3.2.2), different pruning thresholds used to infer complex structural changes, lengths of epochs used to discretize the time span of a data corpus, and the overlap between successive epochs. As explained in the previous section, we compared three inter-topic similarity measures based on the Hellinger distance, the Bhattacharyya coefficient, and what we term the quasi-Jaccard similarity. Different combinations of epoch lengths and successive epoch overlaps are summarized in Tables 1 and 2 for, respectively, the ASD and the metabolic syndrome data sets. Notice that due to the different time spans of the two data sets, different epoch lengths were used in the corresponding experiments. Similarly, epoch overlaps were adjusted (to the closest year) so as to be comparable on a relative basis so that, for example, the maximum overlap examined was approximately half of the duration of an epoch (10:5 vs. 6:3). At the other extreme we also performed experiments using no epoch overlap, as well as an additional setting per experiment with an overlap of approximately 25% of epoch length (10:2 and 5:1 in Table 1, and 6:1 in Table 2).

Table 1 Different combinations of values of the free parameters of the proposed algorithm used for the experiments on our autism spectrum disorder abstracts data set
Table 2 Different combinations of values for the free parameters of the proposed algorithm used in our comparative evaluation experiments on the metabolic syndrome abstracts data set

Similarity measures To begin with, consider the results summarized in Fig. 6. The figure comprises three sets of plots with each set corresponding to one of the three compared inter-topic similarity measures, and showing the number of topic births, deaths, merges, and splits per epoch, normalized by the total number of topics inferred in that epoch. Experiments were performed using three different operating cut-off points on the empirical cumulative density function of inter-topic similarities across the initial temporal graph—the results are shown using lines of different styles, as per the plot legends. The epoch length and the amount of overlap of successive epochs were kept constant in all experiments.

Comparing the corresponding plots across different similarity measures, it can be readily observed that the results obtained using the Hellinger distance and the Bhattacharyya coefficient are very similar. This is unsurprising considering the close relatedness of the two measures, as highlighted previously in Sect. 4.2 and as expressed by the expression in (16). However, at first sight the results obtained using the quasi-Jaccard index appear rather different. This raises the question which similarity measure is better (and on what grounds can such a claim be made given the fundamental lack of objective ground truth), or in a weaker form, which similarity measure should be preferred and when, i.e. if this preference should be universal, application driven, or in some sense intimated by data itself. However, a closer inspection of the plots reveals interesting and important insight that obviates this daunting question. In particular it can be noticed that while there indeed are differences in behaviour between the plots corresponding to the quasi-Jaccard index and those which correspond to the Hellinger distance and the Bhattacharyya coefficient in early epochs, eventually the three converge to approximately the same terminal behaviour. The initial differences are readily explained by the well-known small sample size effect—in the early stages of processing a longitudinal corpus inference relies on much fewer documents than later on, especially in the particular case considered here given that the amount of published research in ASD has been growing rapidly year after year as illustrated in Fig. 7. This hypothesis is further supported by the observation that the choice of different operating points on the CDF of inter-topic similarities across the initial temporal graph, used for its subsequent pruning, also has little effect in the long run—following transient changes, the rates of different types of topic changes converge towards the same pattern.

Fig. 6
figure 6

Comparison of different inter-topic similarity measures using fixed values of the remaining algorithm parameters (epoch length, amount of successive epoch overlap, and cut-off operating points on the similarity CDF used to construct our temporal similarity graph). a Hellinger, 10 year epochs, 5 year overlap, ASD abstracts, b Quasi-Jaccard, 10 year epochs, 5 year overlap, ASD abstracts, c Bhattacharyya, 10 year epochs, 5 year overlap, ASD abstracts

Fig. 7
figure 7

Rapid rise in the rate of publications concerned with ASD. The number of publications per annum indexed by PubMed and matching our criterion for the inclusion in our ASD abstracts data set (please see Sect. 4.1.4)

Fig. 8
figure 8

Effect of the epoch length; ASD abstracts data set. a Hellinger, 5 year epochs, 2 year overlap, ASD abstracts, b Hellinger, 10 year epochs, 5 year overlap, ASD abstracts

Fig. 9
figure 9

Effect of the epoch length; MetS abstracts data set. a Hellinger, 3 year epochs, 1 year overlap, MetS abstracts, b Hellinger, 6 year epochs, 3 year overlap, MetS abstracts

4.3.1.1 Epoch length

We next examined the effects that the choice of the epoch length has on the output of our algorithm. A representative set of results is shown in Figs. 8 and 9. Each figure comprises two sets of plots with each set corresponding to a particular set of algorithm parameters. The key parameter which was varied was the epoch length. For example, in Fig. 8 the results obtained using the epoch length of 5 years was compared with those obtained using the epoch length of 10 years on the ASD abstracts data set. To ensure a fair comparison, successive epoch overlap was not kept constant on an absolute but rather relative basis. Specifically, the overlap is in all cases approximately 50% of the epoch length (5:2 and 10:5 in Fig. 8, and 6:3 and 3:1 in Fig. 9). As before each set of plots displays the number of topic births, deaths, merges, and splits per epoch, normalized by the total number of topics inferred in that epoch, with the three different operating cut-off points on the empirical cumulative density function of inter-topic similarities across the initial temporal graph shown using lines of different styles, as per the plot legends.

4.3.1.2 Successive epoch overlap

Lastly we analysed the effect that the choice of successive epoch overlap has on the output of our algorithm. As in the preceding experiments we visualize a summary of the results using sets of plots of normalized rates of topic birth, death, splitting, and merging, each set corresponding to a different overlap—please see Fig. 10. Even a cursory examination of the plots readily reveals that the parameter in question has a profound effect. While a consistent pattern of changes can be observed in each set of plots, the most noticeable effect is on the rate of topic death. In particular, it can be seen that as epoch overlap is reduced, the rate at which topics die off is increased. For example on our ASD data set using 10 year epochs, for the overlap of 5 years and the similarity CDF operating point \(\zeta =0.95\) the steady state topic death rate is approximately 0.3. It increases to approximately 0.5 as the overlap is reduced to 2 years, and then to over 0.9 for no overlap at all. In other words, in the latter case over 90% of topics exhibit no continuity of any sort (evolution, merging, or splitting), instead disappearing already within the epoch of their birth. Qualitatively this comes as no surprise—the less overlap there is between successive epochs, the less relatedness can be expected between the sets of topics extracted in successive epochs. However, qualitatively, the magnitude of the effect is rather astonishing. Considering that most of the methods described in the literature which discretize time by epochs adopt the no-overlap design, our finding provides strong and valuable evidence that the performance of these methods could be improved with little effort, merely by a slight alteration in the manner discretization is performed.

Fig. 10
figure 10

Effect of the epoch overlap; ASD abstracts data set. a Hellinger, 10 year epochs, no overlap, ASD abstracts, b Hellinger, 10 year epochs, 2 year overlap, ASD abstracts, c Hellinger, 10 year epochs, 5 year overlap, ASD abstracts

4.3.1.3 Parameter combinations and topic life expectancy

In the experiments so far we examined how the topic structure of a longitudinal document corpus and the evolution of this structure over time is affected by different free parameters of the proposed algorithm. Our results suggest that our algorithm is not very sensitive to the exact choice for the value of its parameters. This behaviour is highly desirable because it obviates the need for substantial amounts of data needed to learn sensible parameter values for a particular application. The one parameter which we found to be of particular importance is the amount of overlap between successive epochs used to discretize time. In particular, we found that while the precise amount of overlap is not of particular importance, the introduction of some overlap (25–50% of the epoch length) significantly improves the ability of our algorithm to capture temporal structural changes in the topic content of a corpus. This was most significantly demonstrated in the rate of topic death. As we noted earlier, without any epoch overlap most topics die off within the same epoch in which they are created. Hence we sought to investigate how the topic life expectancy is affected by different combinations of our algorithm’s parameter values.

In the experiments described so far we examined the normalized topic birth and death rates independently. In other words we looked at the proportion of topics which are, respectively, newly created (born) and which disappear (die) as a proportion of the total number of topics extracted in the corresponding epoch. This information does not provide any insight into which topics die off, that is, how long a specific topic has been in existence before it disappears. Here we adopt a more nuanced approach which comprises the following steps:

  • Identification of topic creation: We identify the epoch in which a particular topic was created as the epoch of its birth or the epoch in which the topic is created through the splitting or the merging of topics from the previous epoch.

  • Topic tracking: Following the creation of a topic we track it in the context of complex changes of its topic environment by considering its natural descendent to be its child in our temporal topic graph, with the highest degree of inter-topic similarity across all siblings. A topic is considered extant as long as it has any descendents.

  • Identification of topic death: Finally, we identify the epoch of a topic’s death as the epoch in which the topic no longer has any descendants (children in our temporal topic graph).

Adopting this methodology we analyse the distribution of the life expectancy of a topic across our corpus and the effects of different algorithm parameters. Using the operating point corresponding to \(\zeta =0.95\) on the inter-topic similarity CDF, we performed experiments using the six settings summarized in Table 1. Our results are shown in Table 3.

Lastly, we examined the manner in which topics extracted by our algorithm cease to exist. In particular, for each new topic (where ‘new’ in this context is taken to mean that a topic is either newly born, as defined previously, or that it is created by splitting or merging of topics from the previous epoch) we compute the time until it either dies (i.e. has no offspring in the following epoch), or splits or merges without any of its children having it as the sole parent. The condition that none of a topic’s children have the original topic as the sole parent is important because if there are such children, they can be seen to be evolved successors. If there are multiple children which have the ancestral topic as the sole parent then we track the lifetime of the topic down that path which leads to the longest topic lifetime. Our results are summarized in Table 3 and the plots in Fig. 11. In Table 3 it is important to observe the lack of sensitivity of our method to its specific settings. From the plots in Fig. 11 it is interesting to notice that most of the topics cease to exist already in the first epoch. This observation provides potentially useful insight into the choice of the temporal analysis scale. As expected, increasing the pruning threshold \(\zeta \), which has the effect of decreasing inter-epoch topic linkage, results in an increased proportion of outright topic deaths. Lastly, it is insightful to observe that following the first epoch, for a fixed specific pruning threshold the proportion of topic deaths decreases.

Table 3 Average lifespan of topics (in epochs) extracted from our autism spectrum disorder abstracts data set
Fig. 11
figure 11

A statistical summary of the manner in which topics cease to exist. The plots correspond to our ASD data set, analysed using the Bhattacharyya similarity measure, and a 10 year epochs with a 5 year overlap, and b 5 year epochs with a 2 year overlap

4.3.2 Qualitative results

In the previous section we described an extensive set of experiments providing insight into the role that different free parameters of our algorithm have on its output. Importantly we demonstrated that within a wide range of what may be described as reasonable choices for the parameter values, the proposed method exhibits a high degree of robustness. Following these encouraging quantitative findings, using our (human) higher level semantic understanding of the corpora used, we now examine if the output of our algorithm is meaningful and ultimately useful.

Fig. 12
figure 12

Dynamics of the topic most closely associated with the concept of ‘genetics’. A few dominant words are shown for each topic (shaded boxes)

Case study 1: ASD and genetics

While the exact aetiology of the ASD is still poorly understood, the existence of a significant genetic component is beyond doubt [45]. Work on understanding complex genetic factors affecting the development of autism, which possibly involve multiple genes which interact with each other and the environment, is a major theme of research and as such a good case study on which the usefulness of the proposed method can be illustrated.

We started by identifying the topic of interest as that with the highest probability of the terms ‘gene’ or ‘genetic’ conditioned on the topic, and tracing it back in time to the epoch in which it originated. This led to the discovery of the relevant topic in the epoch spanning the period 1986–1991. Figure 12 shows the evolution of this topic from 1992 revealed by our method (due to space constraints only the most significant parts of the similarity graph are shown; minor changes to the topic before 1992 are also omitted for clarity, as indicated by the dotted line in the figure). Each topic is labelled with its first few dominant terms. The following interpretation of our findings is readily apparent. Firstly, in the period 1992–1997, the topic is rather general in nature. Over time it evolves and splits into topics which concern more specific concepts (recall that such splitting of topics cannot be captured by any of the existing methods). For example, by the epoch 2002–2007 the single original topic has evolved and split into four topics which concern the following:

  • The relationship between mutations in the gene mecp2 (essential for normal functioning of neurone), and mental disorders and epilepsy (it is estimated that one-third of ASD individuals also have epilepsy),

  • Gene alternations, for example, the duplication of 15q11—-13 and deletion of 16p11.2 both of which are associated with ASD,

  • Genetic linkage association analysis and heritability of autism, and

  • Observational work on autistic twins and probands with siblings on the spectrum.

Our framework also allows us to look ‘back’ in time. For example, by examining the topics that the 1992 genetics topic originate from we discovered that the topic evolved from the early concept of ‘infantile ASD’ [39].

Case study 2: ASD and vaccination

For our second case study we chose to examine research on the relationship between ASD development and vaccination. This subject has attracted much attention both in the research community, as well as in the media and the general public. The controversy was created with the publication of the work by Wakefield [62] which reported epidemiological findings linking MMR vaccination and the development of autism and colitis. Despite the full retraction of the article following the discovery that it was fraudulent, and numerous subsequent studies who failed to show the claimed link, a significant portion of the general public remains concerned with the issue.

As in the previous example, we begin by identifying the topic with the highest probability of the terms ‘vaccine’ and ‘vaccination’ conditioned on the topic, and tracing it back to the epoch in which it first emerged. Again, a single topic was readily identified, in the epoch spanning the period 1996–2001. Notice that this is consistent with the publication date of the first relevant publication by Wakefield [62]. The evolution of the topic is illustrated in Fig. 13 in the same way as in the previous section. It can be seen that the original topic concerned the subjects initially brought to attention such as ‘measles’, ‘vaccine’, and ‘autism’. In the subsequent epoch, when the original claim was still thought to have credibility, the topic evolves and splits into numerous others mirroring research directions taken by various researchers. Following this period and the revelations of its fraudulence, the topic assumes mainly single-threaded evolution, at times incorporating various originally separate ideas. For example, observe the independent emergence of the term ‘mercury’. Though initially unrelated to it this topic merges with the topic that concerns vaccination which can be explained by the widely publicized thiomersal (vaccine preservative) controversy (again note that such merging of topics cannot be captured by the existing methods). Although rejected by the medical community due to a lack of evidence, this topic can be seen as persisting to date.

Fig. 13
figure 13

Dynamics of the topic most closely associated with the concept of ‘vaccination’. Notwithstanding the rejection of any link between vaccination and autism, this topic remains active albeit in a form which evolved over time

4.3.2.1 Case study 3: MetS and plasma fatty acids

As noted earlier MetS is highly associated with the risk of developing type 2 diabetes and cardiovascular diseases and is characterized by insulin resistance, abdominal obesity, and high blood pressure, all of which are intimately linked with dyslipidemia and elevated plasma fatty acid levels. In this case study we sought to investigate patterns associated with topics concerning this aspect of MetS.

As in the previous two case studies we began by identifying the topics with the highest probability of the relevant terms (in this case ‘acid’ and ‘fatty’) conditioned on the topic, and tracing it back to the epoch in which it first emerged. For the sake of clarity of visualization we focus on the period starting with the 1999–2002 epoch which shows some of the most interesting dynamics. The evolution of topic of interest in this period is shown in Fig. 14. Examples of illustrative observations based on this section of our similarity graph include the following:

  • The 2001–2004 epoch gives birth to a topic dominated by the terms ‘acid’ and ‘uric’. This topic merges a topic concerned with dietary fat, thereby resulting in a topic associated with hyperuricemia in the following epoch. Both the semantics of the extracted topics themselves as well as the aforementioned dynamics are readily interpreted from and supported by the literature on MetS—the relationship between dietary fat (and types thereof) intake and plasma uric acid levels has been a topic of significant research interest [37].

  • The broad topic concerning fatty acid metabolism present in the 2005–2008 epoch can be seen to branch into three distinct research directions on the accumulation of free fatty acids, controlled clinical trials in murine models, and hyperuricemia and gout. This pattern is again congruent with the actual development of research in the field in recent years: the role of free fatty acid accumulation has been attracting an increasing amount of research attention [52], murine models have been widely used to study specific aspects of MetS in controlled conditions unfeasible with human subjects [44], and there is an accumulating body of evidence supporting a causal link between MetS and the increasing incidence of gout (historically known as ‘the rich man’s disease’) in the Western world [22].

Fig. 14
figure 14

Dynamics of the topic most closely associated with the concept of fatty acids (i.e. in the context of our vocabulary the terms ‘acid’ and ‘fatty’). Fatty acid metabolism plays a key role in the metabolic syndrome

5 Summary and conclusions

This paper focused on the problem of modelling and extracting the topic structure of a longitudinal document corpus over time. The approach we described starts with a discretization of time into epochs which may overlap. Then, using the approximation that the topic structure within each epoch is temporally locally static, the aforesaid structure is modelled and extracted using a hierarchical Dirichlet process. Finally, the evolution of the topic structure over time is captured using a temporal graph underlain by an inter-topic similarity measure. The graph initially populated by edges between all pairs of topics in two consecutive epochs is pruned automatically and the result used to infer complexity structural changes over time which the existing methods in the literature cannot model, including the emergence and disappearance of topics and their evolution over time, as well as the merging and splitting of an arbitrary number of topics.

The proposed framework was evaluated extensively on two large real-world data sets of abstracts of scientific papers, one concerning the autism spectrum disorder and the other the metabolic syndrome. These data were collected by ourselves and made free for public use. Our detailed quantitative analysis of the effects that the free parameters of the proposed method have on its performance revealed a number of important insights. We found that within a wide range of parameter values our algorithm was little affected by the specific value choices. Another important finding, the significance of which extends further than the scope of the proposed algorithm, is that in the discretization of time into epochs it is important that successive epochs overlap. The significantly inferior performance observed with non-overlapping epochs has immediate consequences for the interpretation of previous work and the findings reported in the literature, suggesting a simple and immediate way of enhancing the performance of any algorithm which did not adopt the use of overlapping epochs. Lastly, on several case studies highly relevant to the currently popular directions of research on ASD and MetS, our algorithm’s output was analysed qualitatively and shown to capture well the actual developments in these fields.

While the presented experimental results demonstrate the effectiveness of the proposed framework, our analysis and discussion also highlight several limitations and, thus, promising directions for future research. In particular, our immediate efforts will be directed towards automating the process of model parameter selection, namely the epoch length and overlap (both of which could conceivably vary across time), the temporal graph pruning degree. We expect that this goal can be achieved by learning from large corpora, and optimality criteria based on fundamental principles of information theory for balancing model complexity and expressiveness [7].