EGU Blogs

Research

New open science competition through ARCS

The ARCS (Advancing Research Communication and Scholarship) have launched a new ‘open scholarship’ competition. The aim is simple: describe your ‘open science story’, and you could win $1000! In particular, the competition is aimed at finding success stories – how has practicing any form of open science helped you advance or enhance your career in some way?

All submissions will be published and archived with The Winnower, an open access platform, and assigned a DOI for free so it will become citeable.

To submit your open success story, simply click HERE. Use ARCS2015 as a keyword, so your submission is included in the project collection and considered for a prize. Articles must be submitted before October 5th to be considered for a prize.

[Read More]

The rise of open research data

This was originally posted at: http://exchanges.wiley.com/blog/2015/08/26/the-rise-of-open-research-data/

As a junior researcher in the UK, it has given me great pleasure over the last few years to see the dramatic development of open access publishing. Most major research funders in the UK now require public access to published research articles in one form or another, and many other research intensive nations across the globe are following suit.

Along with this global increase in public access to papers, there has been a gear shift in demand for the availability of additional outputs of research, including code, videos, software, and raw data. One of the most recent steps in increasing access to these outputs has been the RECODE project for researchers in the EU, which seeks to develop an open data ecosystem through shifting research practices. With progress being made in the USA too, the wheels are truly in motion towards a global shift towards open access to all research outputs.

[Read More]

A thought on impact factors

OK, bear with me on this one. It’s a bit of a thought dump, but it would be interesting to see what people think.

You can’t go anywhere in academia these days without hearing about impact factors. An impact factor is a metric assigned to a journal that measures the average number of citations per article over the preceding two year interval. It was originally designed to help libraries select which journals were being used by academics in their research, and therefore which ones they could not renew subscriptions to. However, in modern day academia, it is often used to measure the individual ‘impact’, or quality, of a single paper within a journal – that is, the metric assigned to a journal is used as a proxy for the value of each article inside. It doesn’t make much sense on the face of things, especially when you here stories about how much impact factors are gamed (read: purchased) by journals and their publishers (see link below), to the extent that they are at the least meaningless, and at the worst complete lies.

The evidence suggests that the only thing that an impact factor, and journal rank, is reflective of is academic malpractice – that is, fraud. The higher an impact factor, the higher the probability that there has been data fudging of some sort (or higher probability of detection of such practice). A rather appealing option seems to be to do away with journals altogether, and replace them with an architecture built within universities that basically removes all the negative aspects of assessment of impact factors, at the same time as removing power from profit-driven parasitic publishers. It’s not really too much a stretch of the imagination to do this – for example, Latin America already uses the SciELO platform to publish its research, and is free from the potential negative consequences of the impact factor. University College London also recently established it’s own open access press, the first of its kind in the UK. The Higher Education Funding Council for England (HEFCE) recently released a report about the role of metrics in higher education, finding that the impact factor was too often mis-used or ‘gamed’ by academics, and recommended its discontinuation as a measure of personal assessment. So there is a lot of evidence that we are moving away from a system where impact factors and commercial publishers are dominating the system (although see this post by Zen Faulkes).

But I think there might be a hidden aspect behind impact factors that has often been over-looked, and is difficult to measure. Hear me out.

Impact factors, whether we like it or not, are still used as a proxy for quality. Everyone equates a higher impact factor with a better piece of research. We do it automatically as scientists, irrespective of whether we’ve even read an article. How many times do you hear “Oh, you got an article in Nature – nice one!” I’m not really sure if this means well done for publishing good work, or well done for beating the system and getting published in a glamour magazine. Either way, this is natural now within academia, it’s ingrained into the system (and by system, I include people). The flip side of this is that researchers then, following this practice, submit their research which they perceive to be of ‘higher quality’ (irrespective of any subjective ties or a priori semblance of what this might mean) to higher impact factor journals. The inverse is also true – research which is perceived to be less useful in terms of results, or lower quality, will be sent to lower impact factor journals. Quality in this case can refer to any combination of things – strong conclusions, a good data set, relevance to the field.

Now, I’m not trying to defend the impact factor and it’s use as a personal measure for researchers. But what if there is some qualitative aspect of quality that it is capturing, based on this? Instead of thinking “It’s been published in this journal, therefore it’s high quality”, it’s rethinking it as “This research is high quality, therefore I’m going to submit it to this journal.” Researchers know journals well, and they submit to venues for numerous reasons – among them is the appropriateness of that venue based on its publishing history and subject matter. If a journal publishers hardcore quantitative research, large-scale meta-analyses and the sort, then it’s probably going to accrue more citations because it’s of more ‘use’ – more applicable to a wider range of subjects or projects.

For example, in my field, Palaeontology, research typically published in high impact factor journals involves fairly ground-breaking new studies regarding developmental biology, macroevolution, extinctions – large-scale patterns that offer great insight into the history of life on Earth. On the other hand, those published in lower impact factor journals might be more technical and specialist, or perhaps regarding descriptive taxonomy or systematics – naming of a new species, for example. An obvious exception to this is anything with feathers, which makes it’s way into Nature, irrespective of it’s actual value in progressing the field (I’ll give you a clue: no-one cares about new feathered dinosaurs any more. Get over it, Nature).

So I’ll leave with a question: do you submit to higher impact factor journals if you think your research is ‘better’ in some way. And following this, do you think that impact factors capture a qualitative aspect of research quality, that you don’t really get if you think about what impact factors mean in a post-publication context? Thoughts below! Feel free to smash this thought to shreds.

Welcome to the Open Glossary

At the London satellite event for Open Con earlier this year, myself and Ross Mounce were given some useful feedback after our joint talk on ‘Open Data’ by one of the attendees. Apparently, some of the terminology was too complex, or specialist, for the subject, and some of the talk was unable to be followed unless you were already an expert in the issue.

Now obviously this is something that, as members of the ‘open community’, we do not want. As we progress to setting the default to open, I want it to be an open cultural movement, rather than an exclusive clique. For this, I believe it is important that the terminology we use is designed to be inclusive, rather than accidentally (or otherwise) creating rifts within academia.

To that end, afterwards in the pub, where all things science occur, we decided to create the Open Glossary. This is a resource designed to equip people with the terminology that is used within discussions about the general field of open scholarship. Additionally, it possesses numerous external resources that may be of use. This has been a crowd-sourced effort (original document), so thanks to everyone who has provided feedback, edits, and comments to date. I expect to update it every few months.

What I ask is for people to host this document, and share as broadly as possible with friends and colleagues. And not just those who are interested in science or already active researchers – awareness of this sort of thing is equally as important in being active about it, in many cases.