EGU Blogs

Altmetrics

Which palaeontology stories in 2015 captured the public’s imagination?

This was originally posted here!

Happy New Year everyone! It’s that time of year when all the summaries of an amazing year of research are coming out, and goodness, what a year it’s been! The folk over at Altmetric have been kind enough to summarise the top 100 articles of 2015, measured by their altmetrics scores – a measure of the social media chatter around articles. All the data are available on Figshare, and here I just wanted to highlight the palaeontology stories that stood out in the media this year according to the list.

[Read More]

A thought on impact factors

OK, bear with me on this one. It’s a bit of a thought dump, but it would be interesting to see what people think.

You can’t go anywhere in academia these days without hearing about impact factors. An impact factor is a metric assigned to a journal that measures the average number of citations per article over the preceding two year interval. It was originally designed to help libraries select which journals were being used by academics in their research, and therefore which ones they could not renew subscriptions to. However, in modern day academia, it is often used to measure the individual ‘impact’, or quality, of a single paper within a journal – that is, the metric assigned to a journal is used as a proxy for the value of each article inside. It doesn’t make much sense on the face of things, especially when you here stories about how much impact factors are gamed (read: purchased) by journals and their publishers (see link below), to the extent that they are at the least meaningless, and at the worst complete lies.

The evidence suggests that the only thing that an impact factor, and journal rank, is reflective of is academic malpractice – that is, fraud. The higher an impact factor, the higher the probability that there has been data fudging of some sort (or higher probability of detection of such practice). A rather appealing option seems to be to do away with journals altogether, and replace them with an architecture built within universities that basically removes all the negative aspects of assessment of impact factors, at the same time as removing power from profit-driven parasitic publishers. It’s not really too much a stretch of the imagination to do this – for example, Latin America already uses the SciELO platform to publish its research, and is free from the potential negative consequences of the impact factor. University College London also recently established it’s own open access press, the first of its kind in the UK. The Higher Education Funding Council for England (HEFCE) recently released a report about the role of metrics in higher education, finding that the impact factor was too often mis-used or ‘gamed’ by academics, and recommended its discontinuation as a measure of personal assessment. So there is a lot of evidence that we are moving away from a system where impact factors and commercial publishers are dominating the system (although see this post by Zen Faulkes).

But I think there might be a hidden aspect behind impact factors that has often been over-looked, and is difficult to measure. Hear me out.

Impact factors, whether we like it or not, are still used as a proxy for quality. Everyone equates a higher impact factor with a better piece of research. We do it automatically as scientists, irrespective of whether we’ve even read an article. How many times do you hear “Oh, you got an article in Nature – nice one!” I’m not really sure if this means well done for publishing good work, or well done for beating the system and getting published in a glamour magazine. Either way, this is natural now within academia, it’s ingrained into the system (and by system, I include people). The flip side of this is that researchers then, following this practice, submit their research which they perceive to be of ‘higher quality’ (irrespective of any subjective ties or a priori semblance of what this might mean) to higher impact factor journals. The inverse is also true – research which is perceived to be less useful in terms of results, or lower quality, will be sent to lower impact factor journals. Quality in this case can refer to any combination of things – strong conclusions, a good data set, relevance to the field.

Now, I’m not trying to defend the impact factor and it’s use as a personal measure for researchers. But what if there is some qualitative aspect of quality that it is capturing, based on this? Instead of thinking “It’s been published in this journal, therefore it’s high quality”, it’s rethinking it as “This research is high quality, therefore I’m going to submit it to this journal.” Researchers know journals well, and they submit to venues for numerous reasons – among them is the appropriateness of that venue based on its publishing history and subject matter. If a journal publishers hardcore quantitative research, large-scale meta-analyses and the sort, then it’s probably going to accrue more citations because it’s of more ‘use’ – more applicable to a wider range of subjects or projects.

For example, in my field, Palaeontology, research typically published in high impact factor journals involves fairly ground-breaking new studies regarding developmental biology, macroevolution, extinctions – large-scale patterns that offer great insight into the history of life on Earth. On the other hand, those published in lower impact factor journals might be more technical and specialist, or perhaps regarding descriptive taxonomy or systematics – naming of a new species, for example. An obvious exception to this is anything with feathers, which makes it’s way into Nature, irrespective of it’s actual value in progressing the field (I’ll give you a clue: no-one cares about new feathered dinosaurs any more. Get over it, Nature).

So I’ll leave with a question: do you submit to higher impact factor journals if you think your research is ‘better’ in some way. And following this, do you think that impact factors capture a qualitative aspect of research quality, that you don’t really get if you think about what impact factors mean in a post-publication context? Thoughts below! Feel free to smash this thought to shreds.

One small step for Nature..

Macmillan have released an interesting press release, announcing that all research papers published in their 49 Nature Publishing Group (NPG) journals, including Nature, will be made free to read online, via one of Digital Science’s pet projects, ReadCube (note that Digital Science is also owned by Macmillan). These articles can be annotated in ReadCube, but not copied, printed, or downloaded.

This is not open access*, and NPG have been very careful and explicit about stating this.

What is the reason for this move, then, when we have a globally shifting environment towards open access? Well, academics love to break rules. We share papers freely, and often illegally, with our colleagues all the time. It’s a sort of passive rebellion against paywall-based publishers. A great example of this is #icanhazpdf on Twitter, whereby articles are requested, and then hopefully shared privately by someone else. This kind of activity is what NPG are calling ‘dark social’, like some terrible name for an evil media organisation. By this, they simply mean sharing, but out of their control. This new initiative seems to be a way of controlling, and legitimising this sort of ‘peer-to-peer’ practice.

[Read More]

The future of scientific publishing

Last night, the Society of Spanish Researchers in the UK, SRUK, hosted an event discussing the past, present and future of scientific publishing (event details). One thing that was nice about this discussion, compared to previous ones I’ve attended in London, was the number of practising academics in the room. Often, academics are excluded from the discussions about scholarly publishing, which is a bit odd when you know, they’re the ones who actually need the services that publishers etc. provide.

Anyway, what did we all discuss?

Three great and varied speakers formed our menu tonight. For starters, we had Cameron Neylon, ex-scientista, and now the Advocacy Director for the megajournal PLOS. The main course consisted of Eva Amsen, also an ex-scientist, current epic science communicator and Outreach Manager for F1000Research. Dessert was the experimental Prof. Juan Aréchega, Professor of Cellular Biology at the University of Basque County, Spain, and Editor for the International Journal of Developmental Biology (note: not a predatory journal). I’ll try and summarise some of their key points.

[Read More]