I haven't written about altmetrics so far. Not because it’s not a worthwhile subject, but because there’s so much I don’t know where to begin.
The term “altmetrics” was first suggested in a Tweet by Jason Priem, a PhD student from The University of North Carolina, Chapel Hill and co-founder, with Heather Piwowar, of Impact Story (disclaimer: Jason is an occasional co-author). It is short for “alternative metrics,” though these metrics are better described as “complementary metrics” or “additional metrics.” So, altmetric is the latest buzzword in a long line of something-metrics terms for Web-based indices.
Three years ago Jason Priem, Dario Taraborelli, Paul Groth and Cameron Neylon authored the Altmetric Manifesto (2010). Its first line is a painful truth: “No one can read everything.” They wrote about the ways our traditional filters – peer-reviewed journals – are lacking. They listed problems with peer-review, existing metrics like the h-index, and the Journal Impact Factor’s many flaws as factors in the decision to look for new indices. What kind of new indices? Well, mostly social-media based ones. Everything from Twitter to Instagram can be a potential source for altmetric data, as long as there’s scholarly material there. Of course that to me research blogs seem the most promising source, but you know I’m biased. Less than three years later, the sub-field is practically exploding with altmetric articles. It’s interesting to note most of them, when trying to prove a source is valid, show its correlation with citations. Citations are still the standard for impact.
The Public Library of Science (PLoS) journals are pioneers among journals in presenting both traditional and alternative metrics for each article. The metrics page for a PLoS article shows everything from usage metrics to Mendeley bookmarks to journal citations. Nature has recently adopted the altmetric “donut” from the commercial firm altmetric.com, which collects data from various altmetric sources and presents them as, well, donuts.
Donuts tend to be pale blue because that is the color which represents Twitter, and there are many more tweets than, say, blog posts. The dark blue is for Facebook, yellow for blogs, etc. Unfortunately, Nature using the altmetric donut in every article has ruined my “altmetric” Google alert and now I keep getting regular Nature articles instead of relevant altmetric ones.
Holbrook, Barr and Brown (2013) wrote a short letter to Nature where they summed up many possibilities for alternative metrics; everything from “Provoking lawsuits” through “Textbook authored” to “Coining of a phrase." I’m not sure about all the metrics in the table – I doubt the authors are – but they represent fascinating possibilities.
Table 1: Other possible indicators of impact
Public engagement | Academic community | Media |
Protests, demonstrations or arrests | Invitations to present, consult or review | Article downloads |
Provoking lawsuits | Interdisciplinary achievements | Website hits |
Angry letters from important people | Adviser appointments | Media mentions |
Meetings with important people | Reputation of close collaborators | Quotes in media |
Participation in public education | Reputation as a team member | Coining of a phrase |
Mention by policy-makers | Textbooks authored | Trending in social media |
Public research discussions | Citation in testimonials and surveys | Blog mentions |
Muckraking | Audience size at talks and meetings | Book sales |
Quotes in policy documents | Developing a useful metric | Buzzword invention |
Rabble rousing | Curriculum input | Social-network contacts |
Engagement with citizens abroad | Faculty recommendations, prizes | Television and radio interviews |
New metrics come with new challenges. One challenge is the reliability of data. Take the problem of name variations; It has have always been one of the banes of the bibliometricians’ existence, and it followed us to alternative metrics. That, of course, is why you should have an ORCID number. An ORCID number to researchers is what the DOI is for articles. No matter the name variation, your number will always stay the same, so all your products will be accredited to you and you’ll have the bibliometricians’ eternal gratitude as well. The new tools for aggregating altmetric data are sometimes less than reliable, and each shows different results(!). Stefanie Haustein, my co-author in two conference articles, showed at the end of her presentation of the second article, and you can see the results for yourself:
Different metrics (taken from Haustein et al.'s presentation)
As Stefanie says, these tools are black boxes.
Another problem is the potential gaming of alternative metrics. Not that citations aren’t “gamed” regularly, but it takes a bit more effort, and you can’t buy citations (perhaps barter them, but usually not buy outright). On the other hand, one can buy today tweets, Facebook “likes” and Mendeley bookmarking. Once publishers, editors and researchers will take notice of alternative metrics an arms race will begin, and it’ll be much harder for bibliomericians to catch fraud, since there are so many data sources.
What do alternative metrics actually tell us? In my opinion, that is the biggest challenge. There are many theories about the meaning of citations, but the general assumption is that citations represent the influences on the authors citing them (gross over-simplification, I know). However, the sources for alternative metrics vary significantly and (we assume) represent different kind of impact. You can’t compare tweets to bookmarks to an f1000 expert opinion to a blog post. Another thing is that relevant sources change. Will Twitter be a relevant source in five years? What about Mendeley? Alternative metrics are dependent on information sources that might not be sustainable.
When will alternative metrics stop being alternative? My answer is that they’ll continue to be “alternative” until you see promotion committees and funding agencies use them regularly.
Something completely different:
If you remember, I wrote a while back about the Leiden ranking. In the mathematics and computer science field there is a weird anomaly: The Turkish university Ege is in the second place, after Stanford and before Harvard. When I first saw it I assumed it was a “Shelx” case, where one or two highly impact articles allowed Ege to reach that ranking. However, a Turkish reader, Tansu Kucukoncu, wrote to me that the Ege ranking might be based on a fraud: out of the 210 publications of Ege university in the field, 65 were written by one person, Ahmet Yildirim, who has written 279 overall. Yildirim has been blamed of plagiarism by the Journal of Mathematical Physics and his article was retracted. If you want to read more about the case, I recommend a post by the Paul Wouters from Leiden, who is involved with their ranking.
Stefanie Haustein, Isabella Peters, Judit Bar-Ilan, Jason Priem, Hadas Shema, & Jens Terliesner (2013). Coverage and adoption of altmetrics sources in the bibliometric
community ISSI conference arXiv: 1304.7300v1