Hirshman and colleagues performed a systematic review of publications by American authors that have reported on the survival effect of gross total surgical resection in patients with high-grade gliomas. They identified 108 scientific papers and now publish their fourth paper based on this review of the American glioma literature [11]. Internet searches were performed to determine the authors’ medical subspecialty as well as the timing and location of their medical school, residency, and fellowship training. Links were drawn between authors if one was a faculty member at an institution while another was a trainee in the same discipline. To be connected, mentor and trainee had to be located at the same institution during the same calendar year. From this genealogic approach, the authors identified different scientific herds and previously reported that articles by genealogy members tend to share similar conclusions and that articles by members of the same genealogy tend to publish their results in certain journals [9]. Thus, we scientists seem to stick to our herds and mentors when it comes to both study conclusions and publishing channels. They also reported that neurosurgical publications more often tend to support maximal resection than publications by radiation oncologists or medical oncologists [10]. In the present article, based on the same review of the same 108 American glioma papers, the authors (or nearly the same authors) now quantify the previously reported impact of genealogy on publication results in some greater detail [11]. However, the difference from their previous work is marginal. Their conclusion is (still) that the academic genealogy of scientific authors is predictive of their findings and this contributes to systematic bias in the published literature. Although Hirschman and colleagues are to be congratulated for their creative and important study approach, it still seems reasonable to point out the irony that the authors seem rather eager to salami publish their own study results while they criticize publication bias in the scientific literature. The authors even argue that “the consistency of the results between our previous and current study builds a strong case for the influence of genealogy on published literature.” Although validation of study findings is utterly important, salami-publishing from the same data material does not strengthen the evidence.

Even so, the authors’ findings are startling and deserve to be debated. Are really neurosurgeons and oncologists reaching different conclusions when they study the same thing, and if so why? Are the effects of surgery underestimated by the oncologists, overestimated by the neurosurgeons, or perhaps both? Do we simply inherit scientific truths from our herds or mentors? Is this for example why surgeons preferring microscopes vs. endoscopes in transsphenoidal surgery reach opposite conclusions in their systematic reviews [2, 12]? Or why awake surgery enthusiasts reach different conclusions in systematic reviews on the impact of awake surgery than others [4, 14]? Or is this why radiotherapy usually reduces recurrence rates in atypical meningioma if studies are done by radiation oncologists [18, 20,21,22] but not necessarily if the studies are done by neurosurgeons [7, 13, 15, 17]?

Is also our interpretation of the evidence different because we belong to different herds or genealogies where perhaps heretics are sanctioned and only accepted facts prevail? Does this make us read the scientific literature like the devil reads the bible? Is this why most vascular neurosurgeons refuse to believe results from the Aruba trial [16] or the Scottish AVM cohort study [1]? Or is this why many spine surgeons still like to fuse degenerative lumbar spines despite the negative trials [5, 6] (under the motto: screw the evidence)?

The present study is an important reminder that not only traditional conflicts of interest matter, but our background, herd, or subspecialty might affect scientific results significantly. However, this may not necessarily always be due to deliberate foul play. When designing studies, our scientific and professional background may affect both why the study was done in the first place, whether a variable is chosen as primary outcome or just included for adjustment in a multivariable model, whether the study was sufficiently powered to find an effect of this variable, and whether the study variable was measured in great detail or just crudely estimated (e.g., volumetric analysis of extents of resection vs. dichotomous gross estimations). Neurosurgeons and oncologists have different perspectives and administer different treatments at different time points. It is therefore perhaps not surprising that when neurosurgeons conduct trials on the effects of surgery in glioma, adjuvant treatment variables may be only crudely assessed or not taken into account [19] while oncologists fail to measure the extent of surgical resection when assessing the effects of adjuvant treatment [3]. It seems likely that this difference in professional focus may affect study findings.

However, if genealogy has such a profound impact, can we still trust the scientific literature? Are we just preaching to our choirs in parallel scientific genealogies? Nietzsche claimed that “there are no facts, only interpretations.” Feyerabend, the Austrian science philosopher who is famous for his anarchistic view of science, went further. He claimed that modern science has deteriorated into a “stupid religion” where ideas by scientists claiming to represent the truth are held in high regards comparable to how the views of bishops and cardinals were accepted centuries ago. There are today a vast number of scientific journals, scientific boards, special interest groups, and conferences and there may perhaps be a congregation with “alternative facts” for everyone? A more attractive thought (or “alternative fact”) is perhaps that the Hirshman studies could be wrong. (Unless you are in their genealogy, this could work.) Hirshman and colleagues have for example previously reported (based on the same material again) that year of publication is associated with study results as the ratio of articles supporting the survival benefit of surgery has steadily increased over time [8]. This might perhaps have to do with better study methods over time, better surgical tools, better and more widespread postoperative imaging, and advances in adjuvant treatment (synergic effects). From the current publication (Fig. 4 in the paper), one may calculate that median year of publication is significantly different between articles by the two studied genealogies, neurosurgeons (A) and radiation oncologists (B) (A = 2012 vs. B = 1996, p = 0.016; the Mann-Whitney U). This confounder was seemingly not adjusted for. We may therefore not have to deny science just yet. Quod erat demonstrandum (Q.E.D). What a relief.