Large scale research infrastructures, so called LSRIs, are sizeable national and international investments. Though there is no unified definition of the term LSRI, what they have in common is that they often require investment at the national or international level. They also require extensive planning, time, and expertise to build and operate.
The relatively small country of Sweden hosts a freshly upgraded national facility called the MAX IV Laboratory, the first worldwide realisation of a fourth-generation light source. This type of facility uses intense beams of light generated when electrons are accelerated near the speed of light. The fourth generation represents a big step in the power, brightness and quality of light that can be squeezed out of the electrons. This technology, which was pioneered in Sweden at the predecessor MAX-lab, is now being replicated all over the world.
In a synergistic move, Lund was also chosen as the site for the European Spallation Source (ESS). This is an internationally funded, multi-disciplinary research facility based on what will become the world’s most powerful neutron source when it opens. The facility also uses an accelerator. In this case, the accelerator is for protons, particles which live in the nucleus of atoms. Smashing these protons into atoms of a particular element releases powerful beams of neutrons, which are the other main resident in atomic nuclei. Neutrons have no charge and can therefore travel deep into matter to probe it.
Neutrons and light are highly complementary when it comes to investigating matter, since they can “see“ through it and interact with it in very different ways. These facilities are thus jointly capable of giving researchers a whole range of options and approaches when it comes to solving their questions. This is why they are worth the cost.
But what is that cost? The price tag for the construction of MAX IV was about 6 billion Swedish krona (about $711 million in 2015) and the corresponding bill for ESS 2014-2027 is €3,3 billion in 2013 Euros (about 28 billion Swedish krona, or $4,4 billion in 2013). Given this massive investment it is reasonable to ask: What is the return in terms of scientific and societal impact? What we are presenting here are first results of a journey to answer this question using an approach based on the historic published output enabled by these facilities.
The questions
There have been numerous studies and reports focused on this topic already, based around specific LSRIs or LSRI groupings. However, as far as we’re aware, none have so far been successful in systematically gathering all publications stemming from the global facilities and their user programs.
In 2022 the opportunity presented itself for us to attempt to do this, and for a good cause. We work in a department at Lund University where support and insight is often needed on demand and fast in order for leadership at all levels to make strategic decisions. To this end, we have access to several tools and data-gathering capabilities. What was unique about this problem was that the data set we would need to look at is not easily accessible in these tools.
The good cause in question was the recently-started national SPIRIT project. The project’s goal is to suggest how Sweden should shape a national innovation and collaboration arena between academia and industry connected to LSRIs. We had already established that most of the national competence needed to use LSRIs was based in academic environments. Since Martin also works as a strategic advisor for LINXS - an environment in which it is highly relevant to keep tabs on these questions - we decided to roll up our metaphorical sleeves and do the work. The key questions we set out to answer were:
We were of course also interested in the development over time and any further insights we could gain from answering the above. In discussion with the people working in SPIRIT, we decided to approach this in the simplest way possible, by collecting the corpus of publications that are connected to use- or development of x-ray, and neutron based LSRIs, and to provide some results from analysing the collected corpus as an in-kind contribution to the SPIRIT project. Papers have authors with their affiliations, so when we have a validated list of publications, we can map the participation by country, organisation, or people involved.
Getting the information
If you want to find out how an organisation like Stanford or other universities are doing in terms of their published output, you can easily select it as a corpus of publications in the tools and databases typically used by universities. At Lund University we have the luxury of choice between two of the largest such databases, Scopus, and Web of Science. However, if we want to look at the set of publications stemming from the use of LSRIs, we have no choice but to go to the websites of each of them and gather the publications by hand. This is because these facilities run User programmes and it is not necessarily the case that the facility itself is among the affiliations on publications resulting from experiments there.
There are other strategies that have been tried, such as searching for keywords or using acknowledgements, but they all have their flaws. Keyword searches need lots of fine tuning, and one can never be sure that the hit rate is good enough to capture everything. Both the false positives and negatives can have error bars that are very difficult to estimate. As for acknowledging LSRIs, in many cases users unfortunately fail to do this correctly or consistently when they publish, even though this is a requirement for access. The success in automated reading and classification of acknowledged entities in publications also varies, and is much less developed than that for authorship and affiliations.
For these and other reasons, a good old collection by hand - “from the horse’s mouth” so to speak (even if we’re talking about some massive, billion-euro horses) - was judged to be best. The strength of this approach is that most LSRIs keep track of and report, openly and transparently, which publications they can say their research environment has contributed to. They do this not least because it is often a requirement in the reporting to their various funding bodies, but also because most would emphatically agree that it is the right thing to do. Below is an illustration of the facilities included.