Pages

Sunday, 10 July 2016

A 1982 paper: " Peer-review practices of psychological journals: The fate of published articles"

I have just come across this.
Petersa, DP & Cecia, SJ (1982) Peer-review practices of psychological journals: The fate of published articles, submitted again. Behavioral and Brain Science, 5(2):187- 195. http://dx.doi.org/10.1017/S0140525X00011183

Abstract

A growing interest in and concern about the adequacy and fairness of modern peer-review practices in publication and funding are apparent across a wide range of scientific disciplines. Although questions about reliability, accountability, reviewer bias, and competence have been raised, there has been very little direct research on these variables.
The present investigation was an attempt to study the peer-review process directly, in the natural setting of actual journal referee evaluations of submitted manuscripts. As test materials we selected 12 already published research articles by investigators from prestigious and highly productive American psychology departments, one article from each of 12 highly regarded and widely read American psychology journals with high rejection rates (80%) and nonblind refereeing practices.
With fictitious names and institutions substituted for the original ones (e.g., Tri-Valley Center for Human Potential), the altered manuscripts were formally resubmitted to the journals that had originally refereed and published them 18 to 32 months earlier. Of the sample of 38 editors and reviewers, only three (8%) detected the resubmissions. This result allowed nine of the 12 articles to continue through the review process to receive an actual evaluation: eight of the nine were rejected. Sixteen of the 18 referees (89%) recommended against publication and the editors concurred. The grounds for rejection were in many cases described as “serious methodological flaws.” A number of possible interpretations of these data are reviewed and evaluated.
This is followed by an extensive and interesting open-commentary, making a very interesting issue, available at: http://journals.cambridge.org/action/displayIssue?jid=BBS&volumeId=5&seriesId=0&issueId=02

KNOWeSCAPE Workshop on "Identification, location and temporal evolution of topics – data and algorithm – comparison of approaches", Budapest 29&30 August

See http://knowescape.org/event/identification-location-temporal-evolution-topics-data-algorithm-comparison-approaches/ for details.

Coherency Models of Review Judgements

I am at a nice interdisciplinary workshop in Berlin on "Coherency-Based Approaches to Decision Making, Cognition and Communication.

The basic ideas come from Thagard (1989), that human 'reasoning' happens in a way to ensure coherency between beliefs rather than the classical logicist picture of reasoning from evidence to conclusions.  For example, as well as forward inference, backward inference from conclusions to the evaluation of evidence is common. This theory has now a reasonable amount of evidence to support it and has been extended to include emotions and goals (Thagard 2006)

As well as Thagard there was a nice talk by Dan Simon of University of South Carolina, explaining many identified 'biases' in human reasoning using the coherency framework. If you want details see his paper on SSRN at http://papers.ssrn.com/sol3/papers.cfm?abstract_id=439984. One of the biases he talked about was confirmation bias and cited a nice paper that I did not know about about Peer Review.  This is:
Mahoney, M (1977) Publication prejudices: An experimental study of confirmatory bias in the peer review system. Cognitive Therapy and Research, 1(2):161–175. http://link.springer.com/article/10.1007/BF01173636
 In this the author sent papers to a selection of reviewers where it was known whether the reviewer agreed or disagreed with the conclusion of the paper (on a controversial issue). The results were that the judgement of the reviewers on the quality of the paper were substantially affected by whether the reviewer agreed with the conclusion.

The coherency model of thought seems to be a good basis for modelling the judgement of reviewers within simulations of the peer review system.


Friday, 8 July 2016

CfP special issue on "Academic Misconduct & Misrepresentation: From Fraud and Plagiarism to Fake Peer Reviews, Citation Rings, Gaming Rankings, Dodgy Journals, “Vacation” Conferences, and Beyond"

in the journal: Research Policy

Edited by: Mario Biagioli and Martin Kenney

This Special Issue solicits social scientific articles examining not only traditional forms of misconduct, but also modalities of misconduct that are meant to “game” the modern metrics-based regimes of academic evaluation.  If traditional misconduct – fabrication, falsification, and plagiarism – concerned fraudulent ways to produce scholarly publications, much of the new misconduct targets the publication system itself, for example by producing fake peer reviews, citation rings among authors and journals, or by publishing articles in dubious journals. 

...

See http://www.journals.elsevier.com/research-policy/call-for-papers/academic-misconduct-misrepresentation-from-fraud-and-plagiar for more details.

Thursday, 30 June 2016

Nature paper: Interdisciplinary research has consistently lower funding success

 See http://www.nature.com/nature/journal/v534/n7609/full/nature18315.html
"...we show that the greater the degree of interdisciplinarity, the lower the probability of being funded. The negative impact of interdisciplinarity is significant even when number of collaborators, primary research field and type of institution are taken into account."

x axis is the proportion of applications funded (Australian research council) and the y-axis is the estimated bias towards interdisciplinary proposals (estimated using a logit regression model).

Environmental, physical, social science and economics were (unsurprisingly) most averse to interdisciplinary proposals.

Thursday, 3 March 2016

Repost: Why publishing negative findings is hard

From retraction watch, the story of a professor who tried to submit his reanalysis of two papers to journals and the obstacles he faced.

At: http://retractionwatch.com/2016/02/17/why-publishing-negative-findings-is-hard/

Abstract of Paper: Job Insecurity in Academic Research Employment: An Agent-Based Model

Abstract:

This paper presents an agent-based model of fixed-term academic employment in a competitive research funding environment.  The goal of the model is to investigate the effects of job insecurity on research productivity.  Agents may be either established academics who may apply for grants, or postdoctoral researchers who are unable to apply for grants and experience hardship when reaching the end of their fixed-term contracts.  Results show that in general adding fixed-term postdocs to the system produces less total research output than adding half as many permanent academics.  An in-depth sensitivity analysis is performed across postdoc scenarios, and indicates that promoting more postdocs into permanent positions produces significant increases in research output.
More details at:  https://drericsilverman.wordpress.com/2016/03/03/paper-submitted-to-alife-xv/