Identifying publications in questionable journals in the context of performance-based research funding

Abstract:  In this article we discuss the five yearly screenings for publications in questionable journals which have been carried out in the context of the performance-based research funding model in Flanders, Belgium. The Flemish funding model expanded from 2010 onwards, with a comprehensive bibliographic database for research output in the social sciences and humanities. Along with an overview of the procedures followed during the screenings for articles in questionable journals submitted for inclusion in this database, we present a bibliographic analysis of the publications identified. First, we show how the yearly number of publications in questionable journals has evolved over the period 2003–2016. Second, we present a disciplinary classification of the identified journals. In the third part of the results section, three authorship characteristics are discussed: multi-authorship, the seniority–or experience level–of authors in general and of the first author in particular, and the relation of the disciplinary scope of the journal (cognitive classification) with the departmental affiliation of the authors (organizational classification). Our results regarding yearly rates of publications in questionable journals indicate that awareness of the risks of questionable journals does not lead to a turn away from open access in general. The number of publications in open access journals rises every year, while the number of publications in questionable journals decreases from 2012 onwards. We find further that both early career and more senior researchers publish in questionable journals. We show that the average proportion of senior authors contributing to publications in questionable journals is somewhat higher than that for publications in open access journals. In addition, this paper yields insight into the extent to which publications in questionable journals pose a threat to the public and political legitimacy of a performance-based research funding system of a western European region. We include concrete suggestions for those tasked with maintaining bibliographic databases and screening for publications in questionable journals.

 

Professors Receive NSF Grant to Develop Training for Recognizing Predatory Publishing | Texas Tech Today | TTU

“With more open-access journals making research articles free for people to view, some journals are charging authors publication fees to help cover costs. While some journals that do this are still peer-reviewed and credible, others are not and will publish lower quality work strictly for profit. The difference can be hard to tell, even to the most seasoned author….”

Plaudit · Open endorsements from the academic community

“Plaudit links researchers, identified by their ORCID, to research they endorse, identified by its DOI….

Because endorsements are publisher-independent and provided by known and trusted members of the academic community, they provide credibility for valuable research….

Plaudit is built on open infrastructure. We use permanent identifiers from ORCID and DOI, and endorsements are fed into CrossRef Event Data.

We’re open source, community-driven, and not for profit….”

Where Can I Publish? Part 2: Is there a definitive list? – Delta Think

“We set out to examine whether there is a definitive, curated list of journals that researchers can use when deciding on their publication venue. While some offer very good coverage, the short answer appears to be that no one index offers a definitive list.

Across all journals, there seems to be overlap of significant proportions of the mainstream indexes. However, fully OA journals present a more varied landscape. You need to combine multiple lists to round up a comprehensive list of curated fully OA journals.

Our analysis has combined over 100,000 ISSNs across over 65,000 titles and, we think it represents one of the most comprehensive round ups of the coverage of curated lists available….”

How Americans view research and findings| Pew Research Center

“The Pew Research Center survey asked about several factors that could potentially increase – or decrease – trust in research findings and recommendations. The two steps that inspire the most confidence among members of the public are open access to data and an independent review.

A majority of U.S. adults (57%) say they trust scientific research findings more if the researchers make their data publicly available. Another 34% say that makes no difference, and just 8% say they are less apt to trust research findings if the data is released publicly….

People with higher levels of science knowledge are especially likely to say that open access to data and an independent review boost their confidence in research findings. For example, 69% of those with high science knowledge say that having data publicly available makes them trust research findings, versus 40% of those with low science knowledge….”

 

Peter Suber: The largest obstacles to open access are unfamiliarity and misunderstanding of open access itself

I’ve already complained about the slowness of progress. So I can’t pretend to be patient. Nevertheless, we need patience to avoid mistaking slow progress for lack of progress, and I’m sorry to see some friends and allies make this mistake. We need impatience to accelerate progress, and patience to put slow progress in perspective. The rate of OA growth is fast relative to the obstacles, and slow relative to the opportunities.”

Peter Suber: The largest obstacles to open access are unfamiliarity and misunderstanding of open access itself

I’ve already complained about the slowness of progress. So I can’t pretend to be patient. Nevertheless, we need patience to avoid mistaking slow progress for lack of progress, and I’m sorry to see some friends and allies make this mistake. We need impatience to accelerate progress, and patience to put slow progress in perspective. The rate of OA growth is fast relative to the obstacles, and slow relative to the opportunities.”

Are open access journals peer reviewed? – Quora

As of today, the Directory of Open Access Journals (DOAJ) lists 13,229 peer-reviewed open-access (OA) journals.

DOAJ deliberately limits its coverage to the peer-reviewed variety, and evaluates each listed journal individually.

At the same time, some scam or “predatory” OA journals claim to perform peer review but do not. They give OA a bad name, and get wide publicity, creating the false impression that all or most OA journals are scams.

Analogy: Some police are corrupt, and cases of (actual or suspected) police corruption get wide publicity. But that doesn’t mean that all or most police are corrupt….”

Blacklisting or Whitelisting? Deterring Faculty in Developing Countries from Publishing in Substandard Journals

Abstract:  A thriving black-market economy of scam scholarly publishing, typically referred to as ‘predatory publishing,’ threatens the quality of scientific literature globally. The scammers publish research with minimal or no peer review and are motivated by article processing charges and not the advancement of scholarship. Authors involved in this scam are either duped or willingly taking advantage of the low rejection rates and quick publication process. Geographic analysis of the origin of predatory journal articles indicates that they predominantly come from developing countries. Consequently, most universities in developing countries operate blacklists of deceptive journals to deter faculty from submitting to predatory publishers. The present article discusses blacklisting and, conversely, whitelisting of legitimate journals as options of deterrence. Specifically, the article provides a critical evaluation of the two approaches by explaining how they work and comparing their pros and cons to inform a decision about which is the better deterrent.