Author ranking system: ‘Impact factor’ of the last author.

AuthorshipWe all know that there is a very little room at the first author position on any scientific paper.  There can only be one name.  Even if two researchers equally contributed to the paper, only one name will appear at the front end of the author list.  According to the current convention, the other equal contributing author cannot put his name at front even on his own resume.  That’s a bummer!

Consider another scenario;  a young researcher who is the major contributor to the paper is on the way to become an independent researcher.  He writes the manuscript and has to decide the author list.  Whom should he put as first author? And the last author? Although there are collaborating scientists, their contribution is too small to grant first authorship.  In this case, the researcher takes the first authorship and also declares himself the corresponding author.  Problem solved!  Not exactly!  This researcher just lost a major point in becoming an expert in his field.

Both these cases illustrate an existing problem of author ranking in a paper.  It is a lesser known fact of scientific publication that funding agencies (including NIH), journals, and often the hiring authorities use softwares to rank the ‘impact factor’ of authors in a publication.  NIH uses such softwares to determine who are the experts in a research field.  These ‘experts’ are then invited to the study sections for reviewing grant applications.  Journals use these softwares to decide who could be potential reviewers for the manuscripts.

On the surface, the idea sounds reasonable.  However, there is a serious flaw in this reliance on softwares to select ‘experts’.  These softwares are mostly primitive and are not designed to rank contribution in multi-author papers.  They are highly biased towards the ‘senior author’  which they determine only by one criterion- the last position on the author list.  Selecting experts based on such faulty method may have ridiculous consequences.

Recently, a well established journal requested a newly minted postdoc to review a research manuscript.  The postdoc was thrilled by this opportunity and took the challenge.  However, we learnt that the scope and content of the manuscript was clearly beyond his expertise.  I don’t know what happened to the manuscript but I am glad to think that there are safeguards against such anomalies.  I must clarify that I am not against inviting new researchers to participate and contribute in the functioning of the scientific community.  However, this should be done with a deliberate choice by program officers and journal editors. It should not happen by mistake. Otherwise it will erode the confidence in validity of the process.

In case you are curious, a current ranking system used by the NIH, for example, gives highest score to the author whose name appears last on a paper.  The software considers the last author as senior author.  The next highest score goes to the first author.  Finally, it does not matter where your name is between the first and the last author, the software assigns you the same low score for ‘contributing authors’.

I see an irony here.  Traditionally, the last author is the senior author who directs the project and in most cases provides funding and laboratory space for the scientific work.  If you want to find out the experts, let common sense prevail- a simple Pubmed search should suffice.  Why do we need technological voodoo to assign complex scoring system to discover the known?


3 thoughts on “Author ranking system: ‘Impact factor’ of the last author.

  1. Nice article, I am wondering if you considered about H-Index or G-Index which essentially talks about researchers productivity in terms of publication citations. Or, do you have any good suggested approach, I am interested to know because I am doing research into this area these days.

    • Burhan:
      Thank you for your comment. As you might have figured that ‘ranking’ a researcher or research is a tricky process. Citation of a research paper is not a measure of its ‘goodness’. There have been many research areas that languished for years or decades before it became apparent that they should be pursued by other researchers. The most famous case would be of Barbara McClintock’s ‘jumping genes’. More recently, sudden bloom in mitochondria research is making newly transformed mitochondria researchers to dig deep into the archives of so called ‘obscure papers’ of the 50’s and 60’s. Therefore, in my view, the whole idea of ranking research is both idiotic and elitist.

  2. Although an old post, I came across this somehow and would like to share a bit here. I completely agree with your views on the author ranking system. At the most these rankings put forth by some crazy algorithms might give a hint of a ‘productivity’ of a researcher. And the ‘productivity’ as we see in recent years is blatantly limited to “publication ability’. The more one publishes in high impact journals, the more his chances of being called a ‘successful’ researcher. Over the years this has resulted in publication oriented research and people are missing a broader perspective and have limited their creativity to fit some data into a flashy journal. The problem here is those who have a fair amount of knowledge in medium scale universities are driven by fear of failure (to publish) rather than a free sense and attitude to ‘discover’.

Leave a Reply

Fill in your details below or click an icon to log in: Logo

You are commenting using your account. Log Out /  Change )

Google photo

You are commenting using your Google account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s