MyCites: a proposal to mark and report inaccurate citations in scholarly publications
Research Integrity and Peer Review volume 5, Article number: 13 (2020)
Inaccurate citations are erroneous quotations or instances of paraphrasing of previously published material that mislead readers about the claims of the cited source. They are often unaddressed due to underreporting, the inability of peer reviewers and editors to detect them, and editors’ reluctance to publish corrections about them. In this paper, we propose a new tool that could be used to tackle their circulation.
We provide a review of available data about inaccurate citations and analytically explore current ways of reporting and dealing with these inaccuracies. Consequently, we make a distinction between publication (i.e., first occurrence) and circulation (i.e., reuse) of inaccurate citations. Sloppy reading of published items, literature ambiguity and insufficient quality control in the editorial process are identified as factors that contribute to the publication of inaccurate citations. However, reiteration or copy-pasting without checking the validity of citations, paralleled with lack of resources/motivation to report/correct inaccurate citations contribute to their circulation.
Results and discussion
We propose the development of an online annotation tool called “MyCites” as means with which to mark and map inaccurate citations. This tool allows ORCID users to annotate citations and alert authors (of the cited and citing articles) and also editors of journals where inaccurate citations are published. Each marked citation would travel with the digital version of the document (persistent identifiers) and be visible on websites that host peer-reviewed articles (journals’ websites, Pubmed, etc.). In the future development of MyCites, challenges such as the conditions of correct/incorrect-ness and parties that should adjudicate that, and, the issue of dealing with incorrect reports need to be addressed.
Within all areas of research, the use of previously published material is an essential building block of knowledge-production . This is so indispensable that, famously, Isaac Newton described the process of discovering the truth by relying on previous explorations as standing on the shoulders of giants . However, research results are sometimes used sloppily and cited inaccurately.
Inaccurate representations of previously published material can mislead readers about the claims of the cited source, and distract the accurate flow of information and history of ideas. This pertains not just to bibliographic errors such as spelling mistakes in authors’ names or an incorrect publication year, but to erroneous quotations and misleading paraphrases. While both bibliographic errors and misleading quotations/paraphrases are problematic, this paper only focuses on the latter issue and refers to this phenomenon with the term inaccurate citation.
Inaccurate citations, their scale and the parties involved
Several studies have gauged inaccuracies in citations within a particular field. For instance, Todd et al. analysed 306 citations in the field of Ecology and concluded that in 11.1% of the citations, “the cited article has been interpreted one way, but could also be interpreted in other ways – including the opposite point” . In 7.2% of the citations, however, “the cited article did not in any way substantiate the assertion or results attributed to it” . In their meta-analysis, Jergas and Baethge review the results of 28 studies - published between 1985 and 2013 - on inaccurate citations in peer-reviewed medical journal articles. Reporting on the analysis of 7321 citations, the worrisome conclusion is that 11.9% of citations have major errors (not at all in accordance with what the cited authors claimed), and an additional 11.5% have minor errors (“inconsistencies and factual errors not severe enough” to contradict a statement by the cited authors) . Most of the 28 studies considered for meta-analysis chose a sample that involved citations of multiple sources. Studies that focus on citations of a single target source yield much more troublesome results.
Authors who manually checked how their publications are being cited and used in other studies report much higher rates of inaccuracies and provide classical examples of inaccurate citations for our analysis. In 2018, for instance, Stang et al. analysed a random sample of 100 publications that cited a 2010 paper that Stang had published about the Newcastle-Ottawa scale (a scale used to judge the quality of observational studies in systematic reviews). While Stang (2010) criticises this scale , 94 out of 100 randomly selected articles that cited Stang’s study claimed that the 2010 paper supported the use of this scale . In a similar exercise, Glenton and Carlsen assessed 205 articles that cited their 2011 paper about sample sizes in focus groups . They found that in 50.7% of citing articles, their descriptive report about typical sample sizes was being used as a normative justification for the sample size in studies that cited it . According to the definition used in Jergas & Baethge’s meta-analysis, both mentioned examples demonstrate a worrisome percentage of major quotation errors. Authors of the above-mentioned studies suspect that their articles were poorly read or not read at all, which echoes the claim that some researchers do not read what they cite and merely copy and paste claims and citations from other papers [9,10,11].
Considering other contributing factors, cited authors may also be partially responsible for inaccurate citation of their works through their biased reporting of results in abstracts [12, 13] or biased use of language that may have confused readers [14, 15]. Furthermore, neither peer-reviewers nor editors are always able to prevent the publication of articles with inaccurate citations; an aspect mentioned by authors of both studies who checked how they had been cited [6, 8]. It may even be unreasonable to expect peer-reviewers who mostly provide their service for free, and editors who often have to manage a massive manuscript flow and are rarely well-compensated, to check every single citation for its accuracy. Given the challenges of spotting all inaccuracies during the editorial process in advance of publications, we believe it is useful to make a conceptual distinction between the publication and the circulation of articles containing inaccurate citations. This distinction allows us to analyse the role of more parties and consider new solutions.
The publication of an inaccurate citation refers to the first time that a resource is referenced inaccurately (Case zero). The circulation of an inaccurate citation, however, pertains to the propagation of the already existing inaccurate citations. This is done by researchers who take the inaccurate interpretation/assertation of Case zero for granted, and without a proper reading of the original item, reiterate, rephrase, or simply copy-paste inaccurate citations.
Whilst it is difficult to completely prevent the publication of papers with inaccurate citations, it is worthwhile to focus on tackling their continued circulation. Shifting the focus from the publication to the circulation allows us to consider the role played by a range of parties that could prevent the propagation of inaccurate citations, but do not do this. While it is true that researchers who copy-past inaccurate citations without reading the original reference make a major contribution to the circulation of inaccurate citations, other parties may contribute to this problem with their inaction (see green boxes in Fig. 1). Some of these parties include (1) readers who notice inaccurate citations, (2) authors who are being cited inaccurately and (3) journal editors.
Informed readers are increasingly expected to be proactive about reporting errors in the literature . However, since commensurate reward structures to incentivise raising concerns about inaccurate citations have not emerged, this is not common practice.
It is reasonable to argue that among the post-publication responsibilities of authors of the cited article, one is to react to inaccurate citations of their work. If authors such as Stang, and, Glenton and Carlsen had not reacted to inaccurate citations of their work, they would have contributed to further circulation of inaccurate citations.
Among the post-publication tasks of editors, one pertains to their response to reported concerns about the soundness of published items . Nevertheless, engaging with reports about inaccurate citations may not always be their top priority. In a famous example, editors of the journal of Experimental Economics noted in an editorial:
“As a general rule, the journal in the future will not publish errata to an article merely to point out omitted, inaccurate, or inappropriate, citations. Exceptions may be made to this rule, but we intend these to be very rare. Those authors who feel that they should have been cited or that were cited inaccurately in an article that we have published will have to use other means, such as posting notices on their own websites or contacting key researchers doing related work directly, to notify the research community of their concerns” .
While it is likely that other factors contribute to publication and circulation of inaccurate citations (e.g., production errors, honest errors, intentional misue of original information or the creation of fake news), to the best of our knowledge, there have not been any large studies exploring them and their importance or relevance in different disciplines. Regarding the particular case of editors, in the absence of qualitative research that explores their willingness to engage with this issue, or quantitative studies that compare the number of reported inaccurate citations against published errata, further analysis of editors’ behaviour cannot be provided. Nevertheless, since inaccurate citations are not among classical instances of misconduct, and, correcting them involves protracted communication between the editor and several parties; one understands why editors might not be keen on engaging with inaccurate citations.
Current ways of reporting inaccurate citations
A summary of current ways of reporting inaccurate citations is presented in Table 1, and described in detail below. To correct inaccurate citations, they must first be located and reported. However, the process of reporting inaccurate citations has major problems that might actually demotivate those who spot them from taking any action and thus allowing them to remain in circulation. Without attempting to be exhaustive, in what follows six options for reporting inaccurate citations, and problems associated with each are introduced.
Contacting the editor of the journal where the publication with an inaccurate citation was published involves finding editors’ contact information and sending an email to them (with the details of the article, page number, paragraph, citation, the cited article, and a description of inaccuracies), all of which is time-consuming. Given the prevalence of inaccurate citations, contacting the editors for each incidence is an inefficient way of reporting errors.
One can submit a commentary/letter about the error to the journal where the paper with an inaccurate citation is published. Readers have been encouraged by their peers for many years to make use of the Correspondence or Comments sections of journals to “point out and correct misleading quotation errors” . Preparing and publishing a letter/comment is time-consuming, both for authors and the editorial team. If the inaccurate citation has not been the linchpin of a paper, this could be considered as a well-intentioned but inordinate act that overburdens the editorial team. It would be extremely inefficient to publish a letter/commentary for every single inaccuracy.
Reporting inaccuracies with an email to the corresponding author of the cited and/or citing article, and allowing them to decide how to proceed is another possibility. Even if this report is not ignored, it is very likely for the reporter to be left outside of the communication loop. In which case, the reporter cannot know what happens with their report and whether it is being acted upon or not. If authors from citing and cited papers get in contact with each other, they may just agree that inaccurate citations should not be repeated in the future and they may never contact the journal or submit a corrigendum.
One can report inaccuracies via social media platforms or other outlets that promote(d) the publication with inaccurate citation. Especially with the help of applications such as Altmetrics, it is possible to find mentions of publications on Twitter, blogs and news outlets. One can post (publicly visible) comments to report inaccuracies. In addition to the possibility of having fake/anonymous/pseudonymous names (that may complicate distinguishing a genuine report from trolls), these platforms often allow account owners to delete unwanted comments. Furthermore, not every researcher is a member of these platforms and given the presence of a wide range of non-experts, laypersons, family and friends, social media is perhaps the worst possible place to report inaccurate citations.
One can make a comment on post-publication platforms such as PubPeer. Although useful for post-publication conversations, PubPeer is not designed for the purpose of reporting inaccurate citations. Prose comments about an article in its entirety, are not linked to citations, which has two drawbacks. Firstly, cited authors will not be notified about the inaccuracy, and secondly, it would not be possible to interrogate the use of a source and notify future readers about an inaccurate citation where it occurs in the body of the paper. Furthermore, PubPeer seems to be moving away from an open and free access model towards a hybrid subscription model, which ristricts users’ access to functionality regarding comments. Their new dashboard that allows creating efficient feedback loops and searching through comments is designed to serve journals and institutions for a price . Another complexity of using PubPeer for this purpose is the possibility of making anonymous comments. While this functionality might be useful at times, it has also contributed to disputes and controversies , and may further complicate managing reports and distinguishing genuine ones from trolls.
Concerns can also be voiced via limited annotation capabilities offered by preprint servers and some journals. The journal of eLIFE, for example, have joined forces with Hypothes.is annotation software and offer annotating capabilities on their website . While innovative and useful, these comments will not be visible to readers who may prefer to read an article on other platforms such as PubMed. Furthermore, these prose annotations are not searchable or linked to citations with digital identifiers. This drawback would not allow having the cited authors notified about the inaccuracy, or generate an index of inaccurate citations.
Towards a new solution
Against the backdrop of the current problems of reporting inaccurate citations, it is pivotal to learn from the shortcomings of current practices in order to simplify and streamline this process. We propose and are piloting a new tool that consolidates annotation capabilities (such as those of Hypothes.is) with persistent identifiers (such as ORCID), as well as Open Citation Identifiers (OCI)  and In-Text Reference Pointer Identifier (InTRePID) , to simplify locating and marking in-text citations. We call this prototype MyCites (see Fig. 2) and welcome correspondence from those who wish to be involved in this endeavour (For further technical explanation about OCI and InTREPiD, and how they can be used to generate a new persistent identifier, see the supplementary file. MyCites is a provisional title for the prototype and the project. We might choose a different name once the tool is developed). We believe that such a tool needs to allow readers, editors and also authors who are cited inaccurately, to receive notifications (see Fig. 3).
Designing a devolved mechanism to annotate inaccurate citations has four advantages. First, by creating a space to raise concerns about inaccurate use of citations, it might deter malpractice. Second, it streamlines the process of highlighting and reporting inaccurate citations. Third, it notifies new readers about inaccurate citations that are spotted by previous readers. Finally, in case it is effectively adopted and used by researchers, it could prevent the continued circulation of inaccurate citations and improve the accuracy of links between citing and cited items.
Allowing ORCID users to (publicly) post personal views about citations is not without challenges, and some of these are mentioned below.
The definition of inaccurcy
It is crucial to develop a theoretical grounding and shed more light on the theoretical/conceptual question of what it means for a citation to be ‘(in)accurate’. It is also necessary to create a taxonomy of the different kinds of inaccurate citations.
Adjudication and arbitration
Given the importance of defining what a correct citation is, identifying one person to have the authority to adjudicate correctness and arbitrate possible disputes could become very complicated. For instance, in cases where the cited paper is co-authored by researchers who may disagree with each other about the best interpretation of their work, someone else would have to decide what interpretation of their work is accurate. Furthermore, in cases where the author(s) are no longer alive, choosing one person to indicate what they meant would be very challenging (especially in Humanities and Arts where one can make alternative interpretations of the text).
Responsibility and accountability
Considerations about responsibility and accountability with regards to inaccurate citations are particularly challenging to address. For example, when inaccuracies result in significant losses (e.g., resources, lives), or, in cases where the cited article makes a biased use of language, or, is poorly written and convoluted, pointing the finger at one person would be challenging. Especially in cases where more than one co-author has been involved in the development of the text, responsibilities are diffused.
Reliability of annotations
It is likely that not every reported inaccuracy is reliable and valid. Addressing this issue in different ways will affect the number of reports and the availability of the tool. For example, one way of ensuring that reported inaccuracies are reliable is to have them confirmed by others (e.g., similar to the peer-review processes where two/three referees are involved). Another approach could be to limit the access of researchers from outside a particular discipline to mitigate the risk of misuse, trolling and wasting people’s time through irrelevant complaints.
Incorrect reports and permanence of comments
There may be compelling reasons for removing/hiding incorrect annotations. In this case, someone (e.g., the cited author or the editor) would have to make this decision. Besides the complexity of indicating one person with authority to remove/hide annotations, this also raises questions about the specific conditions that should be met for an annotation to become removed/hidden.
Legal implications and requirements
Although considerations about the need for curation or moderation of annotations seem more relevant to upkeep and maintenance, in cases that annotations are used to blackmail or defame the competition, legal aspects stand out. Should annotations be immediately live and visible to everyone, or is the use of a vulnerability disclosure model (public disclosure of issues only after a certain period) more acceptable? For example, annotations could be first made visible to main parties such as the citing and cited author as well as journal editors to give parties with major stakes some time to react. How and where will the data be stored, who should own this data and for how long, are among questions that would impact copyrights and the use of comments in the future .
Application to preprint servers and journals that offer annotation capabilities
Preprints are increasingly part of the recognised scientific output. Integrating a new tool into preprint serves, and journals that use other annotation technologies might be challenging. Furthermore, it is not clear whether preprint servers and other journals that allow articles to have a version-number should issue a new version after correcting a citation.
Citation indices, citation identifiers and pointers
This tool needs to be linked with a citation index, and accordingly, will have limitations based on the comprehensiveness and openness of the chosen index (although projects such as Open Citations might gradually resolve this in the future ). Similarly, OCIs and InTRePIDs do not capture all citations yet, which might add further limitations to the coverage. Furthermore, interpreting the link between OCIs, InTRePIDs, and annotations about inaccurate citations might not always be straightforward. For example, if one paper is cited three times (for three different reasons) in another paper, and only one of those citations is inaccurate, a simple index might not be able to capture such complexity.
The indispensability of a new tool
One might question why a new tool is needed at all in the presence of post-publication peer review applications such as PubPeer, and annotation software applications such as Hypothes.is. We believe that although both PubPeer and Hypothes.is have been successful in achieving their goals, neither is designed to tackle the problem of inaccurate citations. Other initiatives focused solely on citations (e.g., Scite.AI that use artificial intelligence to clarify whether a citation provides supporting or contradicting evidence for the cited claim ), might be very effective in detecting bibliographic errors. However, these tools too are not designed to detect incorrect citations that pertain to erroneous quotations or paraphrases and given semantic complexities of identifying these errors; it is not clear whether artificial intelligence should be used for this purpose at all.
Engagement, uptake and impact
Finally, one might ask whether researchers would care enough to engage with such a system and if yes, what would be the real impact? The engagement of scientists with new tools and resources that are meant to improve the integrity of published research and the impact of these initiatives have indeed shown inconsistencies. For instance, even after the launch of Retraction Watch database in 2018 , and the integration of retracted articles into reference management systems such as Zotero , citation of retracted articles has not stopped . Furthermore, while PubPeer is used by experts who are heavily involved in exposing errors, for example, Dr. Elisabeth Bik (Image forensic expert who searches the biomedical literature for inappropriately duplicated or manipulated photographic images), not every author or editor responds to spotted and reported errors . Without misleading ourselves into thinking that a new tool that allows annotating inaccurate citations would not face similar problems, we believe that more research is needed on this topic. In the absence of a final prototype and empirical data about researchers’ expectation and feedback, it is difficult to theorise about the question of engagement and uptake by the research community.
Inaccurate citations are prevalent to a worrisome degree. While the true impact of inaccurate citations and their effect on knowledge production and research waste remain difficult to assess, current methods of reporting them are largely inefficient and problematic. Hence, they keep circulating in the literature. We believe that this vicious cycle could be slowed down, and suggest a solution that, if adopted and used by researchers, could be helpful in containing the circulation of inaccurate citations.
By utilising available capacities of the modern publishing landscape, we are developing MyCites to annotate inaccurate citations. We believe that since the ultimate owners of the problem of inaccurate citations are members of the scientific community, they should be involved from early stages of developing MyCites. Currently, we are setting up a taskforce with two working groups to further develop this tool. One working group focuses specifically on the conceptual issues, and the other on technical development. Hence, we welcome correspondence from those who can help us with conceptual/technical aspects of this endeavour. The two working groups will work in tandem and be coordinated by the chairing board (authors of this article). We also welcome suggestions for partnership and co-development of this tool.
Availability of data and materials
Grafton A. The footnote: a curious history. Cambridge: Harvard University Press; 1999.
Newton, I. Letter from Isaac Newton to Robert Hooke. 1675. Retrieved from HSP Digital Library https://digitallibrary.hsp.org/index.php/Detail/objects/9792. Accessed 10 Dec 2019.
Todd PA, Yeo DCJ, Li D, Ladle RJ. Citing practices in ecology: can we believe our own words? Oikos. 2007. https://doi.org/10.1111/j.0030-1299.2007.15992.x.
Jergas H, Baethge C. Quotation accuracy in medical journal articles—a systematic review and meta-analysis. PeerJ. 2015. https://doi.org/10.7717/peerj.1364.
Stang A. Critical evaluation of the Newcastle-Ottawa scale for the assessment of the quality of nonrandomized studies in meta-analyses. Eur J Epidemiol. 2010. https://doi.org/10.1007/s10654-010-9491-z.
Stang A, Jonas S, Poole C. Case study in major quotation errors: a critical commentary on the Newcastle–Ottawa scale. Eur J Epidemiol. 2018. https://doi.org/10.1007/s10654-018-0443-3.
Carlsen B, Glenton C. What about N? A methodological study of sample-size reporting in focus group studies. BMC Med Res Methodol. 2011. https://doi.org/10.1186/1471-2288-11-26.
Glenton C, Carlsen B. When “normal” becomes normative: a case study of researchers’ quotation errors when referring to a focus group sample size study. Int J Qual Methods. 2019. https://doi.org/10.1177/1609406919841251.
Liang L, Zhong Z, Rousseau R. Scientists’ referencing (mis) behavior revealed by the dissemination network of referencing errors. Scientometrics. 2014. https://doi.org/10.1007/s11192-014-1275-x.
Simkin MV, Roychowdhury VP. Read before you cite! Complex Syst. 2003;14:269–74.
Wetterer JK. Quotation error, citation copying, and ant extinctions in Madeira. Scientometrics. 2006. https://doi.org/10.1556/Scient.67.2006.3.2.
Duyx B, Swaen GMH, Urlings MJE, Bouter LM, Zeegers MP. The strong focus on positive results in abstracts may cause bias in systematic reviews: a case study on abstract reporting bias. Syst Rev. 2019. https://doi.org/10.1186/s13643-019-1082-9.
Vinkers CH, Tijdink JK, Otte WM. Use of positive and negative words in scientific PubMed abstracts between 1974 and 2014: retrospective analysis. BMJ. 2015. https://doi.org/10.1136/bmj.h6467.
Abrams KM, Zimbres T, Carr C. Communicating sensitive scientific issues: the interplay between values, attitudes, and euphemisms in communicating livestock slaughter. Sci Commun. 2015. https://doi.org/10.1177/1075547015588599.
Kueffer C, Larson BMH. Responsible use of language in scientific writing and science communication. BioScience. 2014. https://doi.org/10.1093/biosci/biu084.
Teixeira da Silva JA. The need for post-publication peer review in plant science publishing. Front Plant Sci. 2013. https://doi.org/10.3389/fpls.2013.00485.
Committee on Publication Ethics (COPE). Flowcharts, responding to whistleblowers-concerns raised directly (version one); 2015. https://doi.org/10.24318/cope.2019.2.25.
Cooper DJ, Gangadharan L, Noussair C. Editors’ note regarding citations of other work. Exp Econ. 2017;20(1):276–7. https://doi.org/10.1007/s10683-016-9497-9.
Asai T, Vickers MD. Citation errors—there is still much to be done. Can J Anaesth. 1995. https://doi.org/10.1007/BF03011085.
Retraction Watch (Blog post). New feature aims to draw journals into post-publication comments on PubPeer. 2017. Available at: https://retractionwatch.com/2017/12/11/new-feature-aims-draw-journals-pubpeer-conversations/. Accessed 19 Feb 2020.
Torny D. Pubpeer: vigilante science, journal club or alarm raiser? The controversies over anonymity in post-publication peer review. In: PEERE International Conference on Peer Review. Rome: PEERE COST Network; 2018. Available at: https://halshs.archives-ouvertes.fr/halshs-01700198/document. Accessed 14 Jan 2020.
eLife. eLife enhances open annotation with hypothesis to promote scientific discussion online. 2018. Available at: https://elifesciences.org/for-the-press/81d42f7d/elife-enhances-open-annotation-with-hypothesis-to-promote-scientific-discussion-online. Accessed 10 May 2020.
Peroni S, Shotton D. Open citation identifier: definition. Figshare. 2019. https://doi.org/10.6084/m9.figshare.7127816.v2.
Shotton D, Daquino M, Peroni S. In-text reference pointer identifier: definition. Figshare. 2020. https://doi.org/10.6084/m9.figshare.11674032.
Teixeira da Silva JA. The issue of comment ownership and copyright at PubPeer. J Educ Media Lib Sci. 2018. https://doi.org/10.6120/JoEMLS.201807_55(2).e001.BC.BE.
Initiative for Open Citations (i4oc). About. Available at: https://i4oc.org/. Accessed 1 Aug 2020.
Scite.AI. Learn More. https://web.hypothes.is/about/. Accessed 1 Aug 2020.
Retraction Watch (Blog post). We’re officially launching our database today. Here’s what you need to know. October 2018. Available at: https://retractionwatch.com/2018/10/25/were-officially-launching-our-database-today-heres-what-you-need-to-know/. Accessed 20 July 2020.
Zotero (Blog post). Retracted item notifications with Retraction Watch integration. June 2019. Available at: https://www.zotero.org/blog/retracted-item-notifications/. Accessed 20 July 2020.
Retraction Watch (Blog post). Authors to correct influential Imperial College COVID-19 report after learning it cited a withdrawn preprint. April 2020. Available at: https://retractionwatch.com/2020/04/26/authors-to-correct-influential-imperial-college-covid-19-report-after-learning-it-cited-a-withdrawn-preprint/. Accessed 20 July 2020.
Editage Insights (Website). I have found about 2,000 problematic papers, says Dr. Elisabeth Bik. Available at: https://www.editage.com/insights/i-have-found-about-2000-problematic-papers-says-dr-elisabeth-bik. Accessed 19 Feb 2020.
Fenner M, Peroni S, Shotton D, Hosseini M, Habermann T. Panel discussion: citations, references, and JATS. In Journal Article Tag Suite Conference (JATS-Con) Proceedings 2019. National Center for Biotechnology Information (US). Available at: https://www.ncbi.nlm.nih.gov/books/NBK540819/. Accessed 19 Feb 2020.
Hosseini M. MyCites, a concept for marking inaccurate citations. PIDapalooza Conference, January 2020, Lisbon, Portugal. doi: https://doi.org/10.5281/zenodo.3630311.
The authors wish to thank and acknowledge Mario Malicki for suggestions that improved the structure, and David Shotton for his helpful comments on using OCIs and InTRePIDs. We also wish to thank and acknowledge the valuable feedback offered by the two peer-reviewers. The first author has presented the suggested solution on three occasions: 1. Poster presentation during the 2019 PIDapalooza Conference in Dublin, Ireland. 2. Panel discussion on citations during the 2019 JATS conference in Cambridge, the UK . 3. Oral presentation during the 2020 PIDapalooza Conference in Lisbon, Portugal .
Ethics approval and consent to participate
Consent for publication
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Hosseini, M., Eve, M.P., Gordijn, B. et al. MyCites: a proposal to mark and report inaccurate citations in scholarly publications. Res Integr Peer Rev 5, 13 (2020). https://doi.org/10.1186/s41073-020-00099-8