Introduction

Universities are generally understood as spaces for fostering research and teaching excellence, activities that demand compliance with high standards of scientific integrity (ALLEA, 2017, 2023). However, in recent years a growing number of empirical studies have warned that university researchers are engaging in certain forms of misconduct (Pupovac et al., 2017; Buljan et al., 2018; Felaefel et al., 2018; Hofmann & Holm, 2019; Haven et al., 2019a, b, c; Hofmann et al., 2020; Ljubenković et al., 2021; Palla & Singson, 2022). In response to this problem, the scientific literature is now exploring the extent of research misconduct in different research contexts (Fanelli, 2009; Pupovac & Fanelli, 2015; Pupovac et al., 2017; Xie et al., 2021). One of the dimensions explored in recent studies is the possible proliferation of misconduct in specific universities. Some examples include studies carried out in the University of Oslo (Hofmann & Holm, 2019), the Universities of Stockholm, Oslo and Odense (Hofmann et al., 2020), Vrije Universiteit Amsterdam, University of Amsterdam and the two Amsterdam University Medical Centres (Haven et al., 2019a, b, c), University of Zagreb (Ljubenković et al., 2021), University of Split School of Medicine (Buljan et al., 2018); University of Rijeka (Pupovac et al., 2017), Cairo University, the American University in Cairo (AUC), Suez Canal University in Egypt, RCSI Medical University of Bahrain and Ain Wazein Hospital in Lebanon (Felaefel et al., 2018) and Pondicherry University in India (Palla & Singson, 2022), among others.

Despite the abundant literature in this field, studies addressing the issue in specific Spanish universities are notably lacking. Investigations have been carried out in this context about the possible existence of misconduct as perceived by editors of journals from the fields of communication, education and psychology (Fonseca-Mora et al., 2014). There are also studies on perceived misconduct in specific areas of knowledge, such as philosophy and ethics (Feenstra et al., 2021) or neuropsychology (Olabarrieta-Landa et al., 2017). Similarly, other studies have examined the number of retracted papers in the field of biomedicine (Dal-Ré, 2020; Marco-Cuenca et al., 2019).

Spain is an especially suitable context for such case studies because of its evaluation system, which uses metrics to measure impact quantitatively (Jiménez-Contreras et al., 20022003; Butler, 2004; Hicks, 2012; Derrick et al., 2013; Delgado-López-Cózar et al., 2021; Feenstra & Delgado-López-Cózar, 2023). This system was first introduced in the experimental sciences in 1989 and later extended to other fields, including the humanities (Marini, 2018; Cañibano, 2018). The main evaluation tools in the Spanish research system are the sexenio (recognition of research performance assessed by a committee every six years, and rewarded as a productivity bonus) and acreditación (‘habilitation’ or tenure review process for promotion).Footnote 1 Although the specific criteria vary for the different fields of knowledge, the main feature of both evaluation systems is the assessment of academic merits according to the impact of the journals where researchers publish, especially the results achieved in JCR (some fields also include the option of SJR).Footnote 2 Indeed, the Spanish university context was highlighted in the Leiden manifesto as a paradigmatic case in the use of metrics to evaluate researchers in the social sciences and humanities (Hicks et al., 2015).

These characteristics are particularly relevant, given that the existing literature frequently points to publication and impact-based research evaluation systems as influential drivers of misconduct proliferation (De Vries et al., 2006; Delgado-López-Cózar et al., 2007; John et al., 2012; Martin, 2013; Pupovac et al., 2017; Liao et al., 2018; Maggio et al., 2019; Holtfreter et al., 2019; Aubert Bonn et al., 2019). The Spanish context is also characterised by low public investment in R&D&I (1.4% of GDP in 2021; INE, 2022) and deteriorating working conditions for teaching staff. This is especially evident in the precariousness of teaching and research careers (shortage of posts, delay in access to tenured or permanent positions) (Santos-Ortega et al., 2015). In sum, the combination of factors leading to a prevailing climate of ‘publish or perish’, frequently mentioned in studies of misconduct (e.g., Haven et al., 2019c; Palla & Singson, 2022; Pupovac et al., 2017), becomes particularly relevant in this context.

The above mentioned circumstances make Spanish universities an excellent setting to deepen our understanding of perceptions and concerns about research misconduct. Thus, in this study we address the following questions: Is the proliferation of misconduct perceived similarly to universities in other countries? Are there variations among different contexts? And what are the causes of such misconduct, according to Spanish university researchers?

Based on these questions, the present study seeks to fill a gap in the research misconduct literature by exploring perceptions of misconduct in the academic community of a specific university: Universitat Jaume I, Castelló. Conducting this study in one specific university makes it easier to integrate different areas of knowledge into the analysis, thus expanding on the abundant information already available for branches of the biomedical sciences prevalent in studies of misconduct (Jefferson, 1998; Gilbert & Denison, 2003; Titus et al., 2008; Stretton et al., 2012; Buljan et al., 2018; Hofmann & Holm, 2019). It also allows us to gather information on researchers at different stages in their academic careers, thus broadening the approach followed in studies focussed on specific moments in the academic journey, such as early career researchers (Hofmann & Holm, 2019; Hofmann et al., 2020; Krstić, 2015). Therefore, the ultimate aim of the study is to gain a detailed understanding of the concerns and perceptions of researchers at a university located in Spain, a paradigm in the use of metrics for research evaluation.

Methodology

This quantitative study is based on a survey using both closed and open questions. The choice of Universitat Jaume I as the setting for the research was based on the following criteria: 1) its size reflects the average for small-medium Spanish universities—11.585 degree students in 2022; 34 degrees and 40 master degrees—(SIUVP, 2023), 2) it offers a broad plurality of knowledge areas—Arts and Humanities, Sciences, Social Sciences, Health Sciences and Architecture and Engineering—and 3) as the 59,6% of Spanish universities, UJI has the legal status of public university (Ministerio de Universidades, 2022).

The study population was divided into professional categories as defined by the European Commission, which differentiates between four groups: First Stage Researcher (R1, up to the point of PhD), Recognised Researcher (R2, PhD holders or equivalent who are not yet fully independent), Established Researcher (R3, researchers who have developed a level of independence), Leading Researcher (R4, researchers leading their research area or field). The population includes all researchers with contracts at the institution and all in its areas of knowledge. The knowledge areas were grouped as follows: Arts and Humanities (AH), Health Sciences (HS), Sciences (S), Social and Legal Sciences (SLS) and Engineering and Architecture (EA). The total population is 1,030 researchers. In surveying this population, we aim to reflect the heterogeneity of research career stages and areas of knowledge, which will reveal perceptions that are both specific to these two variables and shared across them. The study therefore has considerable analytical potential in that perceptions can be specified according to each area of knowledge and professional career stage, while at the same time shared perceptions of misconduct in the university can be analysed.

The open questions in the survey allow us to contextualise the responses to closed questions answered previously, and to observe the extent to which they were understood. They also give participants the opportunity to mention elements that do not appear in the survey questions or optional responses. The survey is structured in five blocks of questions on the following themes: a) open access (8), b) gender equality (4), c) knowledge of research ethics (4), d) ethical governance in the institution (6) and e) research misconduct (7); in this paper we focus on the last block. The official languages of the university, Valencian and Spanish, were used in the survey, which was administered through the Qualtrics digital platform. Survey responses were accepted between 5 May and 13 June, 2021. On the same day the survey was launched, an email was sent to the research community encouraging participation. In addition, two reminders were sent out in the following two weeks.

The university’s Vice-Rectorate for Research and Transfer collaborated in the dissemination of the survey, since the data collected were part of an institutional project on the implementation of research ethics.Footnote 3 The research community was informed that the results would also be used to develop a Code of Best Practices in Research and Doctoral Studies (CBPID).Footnote 4

Prior to its launch, the survey underwent an initial testing phase; the draft questionnaire was discussed in a working group comprising 13 members of the university community. The second version of the survey was then sent out for review. The participants in this testing phase had a variety of profiles, both in terms of stages in academic career (4 R4, 5 R3, 1 R2, 1 R1 and 2 research evaluation specialists), and knowledge areas (AH 3, HS 2, S 2, SLS 4, 2 specialists). Gender parity was also sought (six men and seven women), and five people with prominent positions of responsibility in the institution (vice-rectors, etc.) participated alongside six academic researchers who do not hold positions in the university administration and the two specialists. This process was crucial to add other forms of misconduct and also an open question on potential prevention strategies (Q7). Ultimately, it facilitated the validation of the survey.

Questions

The first two questions on misconduct aim to uncover the researchers’ perceptions of the proliferation of misconduct in two spheres of application: at the state level and at the university itself. The questions are as follows:

Q1 Indicate whether, at the national level, you consider there to be a proliferation of any of the following types of misconduct in your area of knowledge as a whole (mark all the options you consider appropriate).

Q2 Indicate whether, at our university, you consider there to be a proliferation of any of the following types of misconduct in your area of knowledge as a whole (mark as many options as you consider appropriate).

The list consists of the following options:

  • Duplicate publication or self-plagiarism

  • Plagiarism

  • Fabrication or falsification of data

  • Use of personal influence

  • Pressure on publishers

  • False authorship (ghost or honorary authorship)

  • Abuse of power over research staff in lower positions

  • False participation in research projects

  • Failure to declare conflict of interest when publishing results or hiring research support staff

  • Lax supervision of doctoral theses

  • Misuse of available research project resources (including for research staff recruitment)

  • Failure to disseminate research results or to commercialise them where appropriate

  • Fraudulent review of research papers or projects (motivated by aversion or affinity)

  • Failure to protect personal data

  • Other forms of misconduct

  •  ⊗ NO misconduct occurs in this area

The concept of proliferation is particularly relevant for these two questions. They are worded in such a way as to highlight not the presence, but the possible proliferation of certain forms of misconduct, as perceived by the respondents. That means that forms of misconducts that are prevalent might appear as less proliferating than others that respondents perceive rapidly increasing in frequency. Another key issue in these questions relates to the types of misconduct listed. Our study offers a wide range of options, purposely going beyond falsification, fabrication and plagiarism (FFP) to include a number of questionable research practices (QRP). The 14 options presented were previously discussed in depth in the working group, and were chosen because their impact could potentially affect all the areas of knowledge in the university. “Other forms of misconduct” allowed respondents to provide additional information. Finally, participants were offered the option “NO misconduct occurs in this area”.

Furthermore, an open question, Q3, “Other forms of misconduct” allowed respondents to provide additional information.

The survey also includes a question on possible causes:

  • Q4. You identified forms of research misconduct in one or both the spheres in Q1 and Q2. What do you consider to be the cause(s) of these forms of misconduct? (Mark as many options as appropriate):

    • Spain’s scientific evaluation policy (accreditations–ANECA, and sexenios–CNEAI)

    • The evaluation policy in our university

    • The lack of training in ethical issues for researchers

    • Researchers’ personal ambitions

    • Other reasons

If a participant marked “Other reasons”, they were given the opportunity to elaborate in the following question:

Q5. You selected “Other reasons”; what, in your view, are the motivations for research misconduct:

The respondents were also asked to assess the weight associated with each cause in the following question:

Q6 To what extent do you consider that each cause mentioned influences the appearance of misconduct? (Participants only see the options they marked previously in Q4) Response options: Never, very little, a little, a lot, totally

This was followed by an open question on possible prevention strategies:

Q7. What strategies do you think could be implemented to help prevent future research misconduct?

This open question allowed respondents to expand the information beyond the limits of closed options and numerical data, enabling them to give a deeper, more comprehensive vision of the topic.

The block ends with another open question (Q8) where participants could add further reflections or comment on any points of the survey.

Survey participation

The survey was begun by 539 people, of whom 505 provided sufficient information for our analysis of research misconduct. The remaining 34 people (12 R1, 3 R2, 15 R3 and 1 R4) did not complete the survey. The final number of responses was 505 out of a total population of 1030, that is, 49.03%. For a 95.5% confidence interval (two sigmas) and p = q, the sampling error is ± 3.18% for the whole sample, assuming simple random sampling. The sampling strategy employed (a self-administered questionnaire was sent online to the researchers’ entire population) inevitably generates a self-selection process in the sample. This might lead to bias in the study, as individuals who respond do so for non-random reasons. The sample distribution by academic career stage and area of knowledge is shown in Tables 1 and 2.

Table 1 Population, Sample, and Proportion of Respondents by Category Overview
Table 2 Response rate by field of knowledge

The high level of participation in the open questions is also noteworthy. We received 32 responses to the question on other forms of misconduct, 40 responses to the question on other causes of misconduct proliferation, 203 responses to the question on how to prevent it, and 76 responses to the final open question (in which the topic with the most comments was research misconduct or research ethics). The length of the content of the qualitative section (a total of 19,992 words) reflects the importance the participants attribute to the topic.

The qualitative data was subjected to a thematic analysis to identify consensus and discursive patterns within the corpus. The process was carried out through an inductive process by two individual researchers. This way, a double manual coding process of content analysis was carried out. This approach enabled the identification of patterns for the perception of misconduct typologies, their potential proliferation, underlying causes, and potential preventive measures. We include some of the most notable and illustrative extracts in this article; respondents are identified by professional category R and area of knowledge.

Results

Perception of misconduct proliferation

Our analysis shows that 71.68% (n = 362) of respondents perceive a proliferation of at least one type of misconduct in their knowledge area at the national level, compared to 21.19% (n = 107) who do not perceive this to be the case; 7.13% (n = 36) did not answer this question. The percentage falls significantly when the question refers to misconduct in the university: 48.91% (n = 247) of respondents perceive the proliferation of at least one type of misconduct (this figure represents the cumulative count of respondents who perceive the proliferation of one or more forms of research misconduct), whereas 34.85% (n = 176) do not perceive any misconduct to be on the increase. In addition, the number of non-responses to this question rises significantly to 16.24% (n = 82). In sum, there are substantial differences between the perceptions of misconduct at the national level and in the respondents’ own university (see Fig. 1).

Fig. 1
figure 1

Perception of misconduct proliferation in the Spanish university system and at Universitat Jaume I. Y-Axis (Vertical): Percentage

Considering only those individuals who believe that research misconduct is proliferating, the distribution of misconduct practices is as follows when comparing the Spanish university system and the Universitat Jaume I (See Fig. 2).

Fig. 2
figure 2

Distribution of the number of proliferating misconduct types perceived in the Spanish university system and at the Universitat Jaume I. Y-Axis (Vertical): Percentage

The perceived proliferation of misconduct is observed as well in the open questions (Q5 and Q8), in which several participants reflect on its proliferation. Some of these responses highlight their proliferation as follows:

These (mal)practices are not only substantial, but are likely to become even more so in the future. R3.SLS.41

The types of research misconduct described are relatively common. R3.EA.15.

Research, not the administration, in my area of knowledge lies somewhere between the jungle and a sharks’ feeding frenzy. R3.EA.18

Another type of response refers to the difference between the possible “sporadic” presence of misconduct as opposed to its “proliferation” (something we will return to in the discussion). One participant specifically notes that:

There are some types of misconduct that do occur in my area of knowledge, as far as I know, but I haven’t marked them because the question asks about “proliferation”. I don’t think they are widespread, but they do occur from time to time. R4.AH.7

Regarding the average number of misconduct types marked as proliferating in their field (excluding responses reporting no perceived misconduct) this is distributed as follows according to gender, professional categories and knowledge areas (Table 3).

Table 3 Comparison of the average proliferating misconduct types perceived by gender, professional category, and knowledge area

There are no significant differences in the number of misconduct types perceived as proliferating by gender. With regard to professional categories, group R1 identifies the highest number of misconduct types, followed by R3 and R2; group R4 perceives a lower proliferation of misconduct and reports a smaller difference between the UJI and the Spanish university system in general. By knowledge area, the highest number of misconduct types is perceived in Arts and Humanities for the Spanish university system as a whole. However, the figure drops significantly for the UJI in this knowledge area, which is relevant since it falls below the average misconduct perceived in Social and Legal Sciences. In Sciences, the knowledge area with the fewest perceived misconduct types, the difference between the Spanish system and the UJI is only 0.27, indicating that the perception of misconduct in the UJI is not only very low, but is also consistent with the perception for the Spanish university system. This assumption is supported by comments in the open section, in which some respondents state that misconduct does not affect the Science knowledge area:

In my field of knowledge, the sciences, work is carried out seriously, professionally, responsibly and with a high level of commitment, so research ethics doesn’t seem to me to be such a relevant issue, or at least not something that needs improving. R3.S.4

As reflected in the excerpt, research misconduct appears to be perceived by science researchers as less on the rise in their field of Science due to the nature of the scientific method. This suggests a prevailing hierarchical vision of areas of knowledge in the sciences. In contrast, the difference between perceptions of the Spanish university system and the Universidad Jaume I reported by Arts and Humanities and Health Sciences respondents is 1.32 and 1.27, respectively, suggesting that these are areas with a high perception of misconduct in Spanish universities in general (in contrast to their own).

Types of misconduct perceived as most proliferating

The types of misconduct respondents most frequently report at the state level are 1) use of personal influence, 2) lax supervision of doctoral theses, and 3) abuse of power over research staff. These are also the most proliferating forms of QRP perceived in the university, although options 2) and 3) are reversed (see Fig. 3).

Fig. 3
figure 3

Perception of misconduct proliferation in the Spanish university system and at the Universitat Jaume I. X-Axis (Horizontal): Percentage

These data show strikingly high percentages of the most common perceived misconduct types, especially at the national level. Lower levels of perceived misconduct are found for variables 14) plagiarism and 15) failure to protect personal data, both at the national level and within the university. Perceptions according to professional category, area and gender are reported in Table 4 below.

Table 4 Perception of misconduct proliferation in Universitat Jaume I by area, professional category, and gender. Bold figures show the most significant values for each gender, professional category, and area

By area, self-plagiarism, plagiarism, pressure on publishers, lax supervision of doctoral theses, fraudulent review of papers and non-compliance with data protection stand out in AH. In S, only data fabrication is of note. In HS, false authorship and misuse of research project resources are most commonly cited. In SLS, the use of personal influence, abuse of power and failure to declare conflicts of interest prevail. Finally, in EA, false participation in research projects and lack of dissemination of results are prominent.

By professional category, the highest perceptions of misconduct proliferation are found in groups R1 and R2. It is indicative that lax supervision of theses and plagiarism are the practices most commonly reported by R4, possibly because they are in a better position in the academic hierarchy to observe these practices more easily.

By gender, men perceive a higher proliferation of misconduct. The most striking differences are found in the use of personal influence, a practice that women report 5.9% more often, and in false authorship, reported 7.4% more often by men than by women.

Moreover, it is significant that in the space provided for participants to describe types of misconduct not included in the list (Q3), they referred to publication in dubious or predatory journals. This idea is associated with “aggressive”, “predatory”, “lax” publication models (seven people referred to these issues). These terms can be interpreted as indicative of an effort to delegitimise the researchers who publish in them and their work. Here, researchers’ choice of journals in which to publish their work assumes an ethical dimension.

A number of participants also used this open question to express concerns about or elaborate on misconduct related to relationships with colleagues. This is noteworthy, as it is observed both in the answers to the closed (Q1 and Q2) and open questions (extracts from Q3 and Q8). In sum, misconduct involving interactions between different academic categories emerges as a central concern among the researchers. These types of misconduct refer to the perpetuation of power relations and networks (and the inequalities resulting from them) between different groups and individuals in the academic setting. The following excerpts illustrate this point:

A conception [of the university] as a pyramid, in which the group leader puts their name to everything that comes out, even if they do not participate, in order to increase their index impact and get on the ‘most cited’ or ‘most published’ list, as this results in greater visibility for the group. R3.EA.20

The hierarchical, medieval model the university is based on, which despite being outdated, continues to be reproduced on a daily basis. R2.AH.1

Hierarchies must be urgently penalised so they are not perpetuated in future generations of researchers. The academic career is hard, but it should not be painfully intolerable in a public, fair and open organisation such as our university aspires to be. R2.SLS.21

The worst type of misconduct in the university is linked to the abuse of power by faculty in senior positions in the hierarchy. R3.SLS.55

There are researchers at the UJI who use mafioso/coercive practices with “inferior” teaching staff. R3.SLS.19

Abuse of power over people coming into the university later. R3.EA.24

Using qualified research staff or those in training without paying them (beyond what could be understood as an internship) under the promise of contracts that, if awarded, do not correspond to the work actually done or required of them. R2.AH.4

Possible causes leading to misconduct and suggestions on how to prevent it

The possible causes behind the proliferation of misconduct identified by the respondents relate to two particular factors: 1) personal ambitions, and 2) Spain’s scientific evaluation system. These reasons are given by 51.5% (n = 260) and 47.3% (n = 239) of the respondents, respectively, although notably, 38.08% (n = 155) identified both causes as key factors in the possible proliferation of misconduct. In turn, 14.9% (n = 75) consider the university’s own evaluation policy to be behind the proliferation of misconduct, and 20.6% (n = 104), lack of knowledge about research ethics issues (see Fig. 4).

Fig. 4
figure 4

Reasons given for misconduct proliferation. Y-Axis (Vertical): Percentage

Likewise, 9.11% (n = 36) suggest other possible reasons, describing in the open question (Q5) issues such as excessive bureaucracy, precariousness, lack of time, lack of effective control systems and, significantly, the hierarchical model of the university. This ties in with the question mentioned in the previous section on the problematisation of relations with colleagues.

With regard to the specific weighting of each cause (Q6), scientific evaluation stands out slightly with 4.24, compared to 4.21 for personal ambitions (Fig. 5).

Fig. 5
figure 5

Average importance attributed to the causes of misconduct proliferation. Y-Axis (Vertical): Mean Value of Likert Scale 1-5

These data are also reported according to gender, professional category, and area in Table 5.

Table 5 Average importance attributed to the causes of misconduct proliferation by gender, professional category and area. Bold figures show the most significant values for each gender, professional category, and area

This table shows the average importance attributed to each reason, taking into account only the responses that mentioned them as a reason for the proliferation of misconduct. The differences by gender are not significant, except for the significantly higher importance women attribute to all causes. Earlier career categories give more importance to personal ambitions than to Spain’s scientific evaluation policy. The importance attributed by both male and female R4 researchers to the lack of training in ethical issues is notable. By area, Arts and Humanities researchers highlight the importance of the university’s own evaluation policy, and the lack of training in ethical issues is also reported by Social and Legal Sciences researchers.

The possible origins of the proliferation of misconduct were also addressed in an open question asking participants for their ideas on the best strategies to prevent it (Q7). Here, many of the responses identify a need to change the system of scientific evaluation (66 respondents from different professional categories and areas of knowledge specifically mention this issue). These statements are broad and complex, and are striking in that they identify different parts of the evaluation system, as we will now explain.

Firstly, some of the respondents recognise a causal relationship between the evaluation system and misconduct proliferation. For example:

In my opinion, the misconduct stems from the evaluation system itself. R3.SLS.11

At the national level, accreditación/sexenios requirements are becoming more and more demanding and lead to misconduct that, if it were not for the high level of frustration in many cases, would otherwise not take place. R3.EA.16

The current research evaluation system engenders some of these types of misconduct. R3.S.16

Some types of misconduct are integral to the system itself, for example quantitative productivity. R4.S.27

Temporary teaching and research staff are running a long-distance race where the objective is to publish at any price. R1.HS.1

Misconduct is due to the fact that research staff feel pressured by the demands of academia to publish at any price to get promoted in their teaching and research career. In the end, that’s what counts most. R3.SLS.15

Secondly, several participants specifically mention certain aspects of the evaluation system as driving the proliferation of misconduct, particularly the central role of bibliometric indicators in measuring academic productivity. It is therefore unsurprising to hear critical voices from different areas of knowledge as follows:

Don’t make promotion of researchers in public universities, who do their research with public funds, dependent on the interests and commercial strategies of the private multinational publishing companies that generate the rankings. R3.SLS.30

An overhaul of hiring criteria and ANECA research evaluation systems is crucial. R4.SLS.36

Rethinking incentive policies to include broader criteria than publication in the JCR; critical reflection on the real meaning of journal impact factors. R4.SLS.2

Don’t consider journal impact factors in CV evaluations in ANECA. R4.SLS.3

Journals’ impact factors cannot be “the golden rule” for evaluating the quality of scientific production. R4.SLS.22

Get rid of the JCR impact factor as the key (or practically the only) element for measuring quality. R4.SLS.9

Eliminate the evaluation of research according to journal impact factors. R2.AH.10

Don’t focus all state-level research on journals’ position in the JCR. R4.S.31

Don’t depend so much on journal impact factors. R2.SLS.2

Stop glorifying the journal’s quartile ranking and explore new quality indicators for academic papers. R3.EA.4

Thirdly, respondents refer to the way the evaluation system is driving a change in direction, manifested in the growing pressure to publish, which is prioritised in research CV evaluations. In sum, they are critical of the drift towards “publish or perish” and associate the demands for constant productivity with the proliferation of misconduct. Some researchers call for:

Less pressure from the administration to publish, publish, publish as the main mechanism to forge a career in the university, or simply to survive there. Most of what is published in some areas of knowledge, even in journals with high impact factors, serves almost exclusively for that purpose: so authors can survive in the university or make a career there, and to feed the whole self-referencing wheel of the academic publishing industry, without having any real social impact. R1.SLS.15

Reduce the pressure on research productivity and keep it at levels that are healthy both for the rigorous construction of knowledge and for people’s lives. R3.SLS.103

Reduce the pressure to publish. R3.S.35.

Remove the pressure to publish a minimum number of publications per year/sexenio. R3.AH.3

Reduce the pressure to publish and focus efforts on more influential articles. R4.S.3

Remove the pressure to only have (lots of) Q1 papers in the JCR. R3.EA.11.

Get rid of publish or perish, especially in accreditation for teaching staff. R2.HS.3

Less pressure to publish in high-impact journals. R2.SLS.19.

The figure of the researcher in this country has become a machine for publishing, with no reflection on what they are researching, [or] what impact it has on society. R3.SLS.70

Change the criteria for the accreditation of teaching staff (ANECA) and sexenios and ensure that publications and projects are not assessed “by weight”. R2.HS.5

In sum, these critical voices call for a remodelling of the evaluation system in order to prevent misconduct. To this end, several participants also state that the key solution lies in providing an alternative definition of what “research quality” means. In the main, these statements advocate reducing the weight of quantitative criteria and combining them with qualitative criteria. This sentiment is illustrated in the following extracts:

Research should not be valued in purely bibliometric terms. What is valued today is having a large number of articles that nobody will ever read and that have not brought about any transfer [of knowledge]. This needs to change. R3.EA.61

Qualitative evaluation of merit by experts; avoid extreme “quantification” of all merits; evaluate coherence throughout the career path. R3.EA.44

The above values and practices (transparency, etc.) cannot be extracted and put in the right context just by assessing written CVs and comparing impact factors. Strategies are needed that focus on the evaluation of the research process, based on the research results. R3.EA.5

An evaluation of research activity that doesn’t just take into account the number of publications and the misunderstood “quality” (or equated with metrics that do not measure quality, such as impact factors). R3.SLS.20

Value the content and quality of scientific production and not the number of articles or publications. R2.SLS.9

Value quality instead of quantity. R3.HS.4

Give greater weight to quality (which is not only the impact but also recognised importance and authority over time) than to the quantity of publications. R3.SLS.44

Accreditation not based on the number of publications, but on the quality and impact of the research. R2.AH.3

Although nuances emerge in the respondents’ understanding of “research quality”, they defend aspects such as transfer, social impact, scientific dissemination or originality. Some also advocate placing greater value on the quality of other tasks such as teaching, which takes second place in the requirements of the academic evaluation and promotion system. Notions of what quality should be are varied, but the responses seem to share a central idea: the importance of preventing the means (high-impact publication) becoming the researcher’s ultimate (and only) end. This idea is clearly illustrated in the following excerpt:

[the solution] would be to go beyond the current short-sighted view to take a broader and more critical look at what “quality” in research is considered to be [...]. In the present system, the indicators initially conceived as a means to an end have become ends in themselves. Researchers fall into the trap of not doing research that is necessary and practically relevant. R3.SLS.78

These responses to the open question provide insights into the survey data and, specifically, the perception of the possible causes of misconduct. According to the respondents, it is mainly the demands of the evaluation systems that seem to contribute negatively to the proliferation of misconduct. The general perception is that the current evaluation system, which demands constant productivity from researchers in the form of articles that rank well in the bibliometric evaluation scales, eventually encourages the less desirable inclinations of researchers in a highly competitive and precarious context. It is striking, however, that in the question on prevention (Q7) few respondents mention the personal dimension and the issue of ambition, which was the most frequently identified cause in the closed question. Only four people refer to it at all, stating that ambition is an individual matter and therefore difficult or impossible to control and change. One participant went so far as to state that “having ambition is not bad, but [some ways of] dealing with it are” (R3.SLS.13).

Other responses refer to the explanations already mentioned that situate the most common misconduct in the context of the power relations in interactions between colleagues. On this point, several respondents propose QPR prevention strategies that recognise the need to 1) foster better dynamics in research groups, 2) penalise those who engage in misconduct (and/or conversely, reward those who promote good practices), 3) promote greater transparency within research groups, 4) give fair recognition to the contributions of less established individuals (especially in publications) and 5) eschew false authorship. Finally, others also advocate promoting greater awareness of good practice in research in general and in research groups in particular, that is, better knowledge of research ethics (in line with responses to Q4).

Discussion

Several empirical studies have highlighted the possible incidence of misconduct in research in general and in universities in particular. It is difficult to know the real extent of misconduct in research, either in general or in specific universities, as studies (including this one) tend to collect data on perceptions. Comparisons between studies should also be made with caution. In this regard, for example, the data may change depending on whether the focus is on FFP, or is broadened out to include more QRPs or limited to fewer, or on how the questions are worded (Godecharle et al., 2018; John et al., 2012; Pupovac & Fanelli, 2015; Xie et al., 2021).

The data from our study show that 71.68% of researchers perceive that at least one form of misconduct is on the rise at the national level, a figure that falls to 48.95% when the question refers to their own university. This is common in studies on the perceptions of misconduct, which tend to underreport misbehaviours, the closer they are to home, and overestimate those of their colleagues (Pupovac & Fanelli, 2015). The general proliferation perceived falls short of that observed in some other studies which, while not focussing specifically on individual universities, offer some noteworthy data. For example, in the context of Nigeria, 96.2% of researchers “believed that one or more forms of scientific misconduct had occurred in their workplace” (Okonta & Rossouw, 2014, 3). However, our data do coincide with other studies in universities such as, for example, that of Felaefel et al., (2018, 72) for three universities in the Middle East, which found that “74.5% reported having knowledge of [some] misbehaviors among […] their colleagues”.Footnote 5

Any comparisons between our analysis and other studies should bear in mind that here we are asking about the possible proliferation of misconducts. We opted to use this term rather than “occurred”, as in Okonta and Rossouw, for example, or the concept of “knowledge” of misconduct used in Felaefel et al.’s study. This conceptual choice in formulating our research problem allows us to define which actions are of most concern to researchers at the specific university in our study.

The most proliferating types of misconduct reported by our participants are use of personal influence, lax supervision, and abuse of power over people in lower positions. These are linked to what we might identify, following de Vries et al. (2006), as “life with colleagues”, which is a concern also noted in responses to the qualitative questions in our survey. The participants state that relationships between colleagues are damaged by the power relations and inequality between staff at different points in their academic careers. This configuration of social relations in the academic setting is explained either by the aim to achieve better results in scientific evaluations or by an unwillingness to comply with the obligations and principles of integrity in relations with people lower down the university hierarchy.

Few studies include data on this type of misconduct, and specifically on the use of personal influence or the abuse of power over people in lower positions. One study in the Spanish context highlighted the use of personal influence to manipulate review processes, which reached 57.5% in the areas of philosophy and ethics (Feenstra et al., 2021). The issue of abuse of power has also been explored in studies such as Martison et al. (2005), who found that 1.4% of their sample engaged in “relationships with students, research subjects or clients that may be interpreted as questionable”. However, this study enquired into the researchers’ own actions. Our research thus extends knowledge on these little-studied QRPs.

Lax supervision of doctoral dissertations is a QRP that has been attracting wider attention among researchers (Gopalakrishna et al., 2022; Haven et al., 2019a). The data of our study supports some questions addressed by previous research. For example, Haven et al. identified this type of misconduct as the most outstanding problem in the analysis of the different disciplines (Biomedicine, Natural sciences, Social sciences and Humanities) present in a sample of universities in Amsterdam (Haven et al., 2019a).Footnote 6

In contrast, when compared with other studies on misconduct related to false (ghost or honorary) authorship, our data on proliferation –15.77% for the national level and 11.13% for the university level– are much lower than those perceived in other universities, such as 63% found for Pondicherry University in India (Palla & Singson, 2022); 55.7% for the University of Rijeka (Pupovac et al., 2017), or between 36 and 46% at the Universities of Stockholm, Oslo and Odense (Hofmann et al., 2020). Therefore, and although this implies a comparison conditioned by the diverse nature of the studies, researchers in our study are less concerned about this aspect than about the interpersonal relationships mentioned above.

To a lesser extent, our study also found the proliferation of some forms of misconduct related to the fraudulent review of research articles and/or projects, failure to disseminate research results, misuse of research project funds and the avoidance of conflicts of interest. Of these forms of misconducts, there is a striking difference in data for fraudulent review, perceived by 20.79% at the national level and 6.36% for the university. These figures stand out both for the difference between the two levels and for the reported proliferation, compared with other studies. For example, in the study by Gopalakrishna et al. (2022), the figure is less than 1% (although this study examines researchers’ own behaviour and not perceptions of their colleagues’). In turn, in Haven et al.’s (2019a) study of universities in Amsterdam, “unfair review” only appears among the top five QPRs for the humanities, but is not present in biomedicine, natural sciences and social sciences.

With regard to FFP, our study shows a lower proliferation than other forms of misconduct. This is not a surprise; other studies have been warning of the need to pay attention not only to these forms of misconduct, but also to “mundane misbehaviour” (de Vries et al., 2006; Haven et al., 2019a), as these are precisely the most prevalent types.

Finally, the possible causes of misconduct proliferation are mainly associated with personal ambitions and the Spanish evaluation system. The weight of the evaluation system is also widely covered in responses to the study’s open questions. Participants point out that the system encourages researchers to increase their productivity in order to obtain positive evaluations for teaching accreditation, applications for positions or projects, etc., regardless of the means or channels they use to do so. This association between misconduct and the evaluation system has been highlighted in the abundant literature in the field (Okonta & Rossouw, 2013; Tijdink et al., 2014; Pupovac et al., 2017; Liao et al., 2018; Felaefel et al., 2018; Maggio et al., 2019). Thus, although it is sometimes difficult to establish direct causal relationships between misconduct and evaluation systems (Fanelli et al., 2015), our study builds on the evidence from researchers that the pressure to obtain results and, above all, the measurement of these results with quantitative criteria based on the accumulation of merit, does not help to foster scientific integrity positively.

Consequently, the results of our study point to the potential adverse impact of the current evaluation systems, echoing the concerns of prominent initiatives such as the San Francisco Declaration on Research Assessment (DORA, 2012) and, more recently, the Coalition for Advancing Research Assessment (COARA, 2022). These initiatives demand the academic community and policymakers to implement reforms that not only base hiring or promotion processes on the evaluation of journal-based metrics. Instead, they advocate for an approach that places a primary emphasis on qualitative assessment (by the peer-review approach) complemented by a responsible use of quantitative indicators (COARA). This way, these initiatives sound a cautionary note against the potential perils of fostering a “publish or perish '' culture and call academics and policy-makers to explore alternative methods that integrate a qualitative evaluation.

Study limitations

Our study is limited to exploring the perceptions of academics at a single Spanish university. An interesting line for future studies would therefore be to compare Spanish universities and contrast their findings with our data. Future empirical research in Spain could also explore some of the specific forms of misconduct observed in our study (especially those for which we found differing trends from those seen in other contexts, such as those linked to fraudulent authorship). However, despite these limitations, we believe that the high participation rate in our study (49.03%), as well as the characteristics of this university (explained in the methodology section), make this case an interesting example to extend understanding of possible trends shared by other Spanish universities. Note also that our study takes an initial approach to the perceptions of various forms of misconduct for a specific university in Spain, which has received little research attention to date.

On the other hand, when exploring potential causes for research misconduct—especially in closed questions— it would be advisable to broaden response options for a more comprehensive understanding of the problem. For example, the qualitative data suggested that relationships with colleagues are becoming a central concern among academics. Future studies would benefit from broadening response options to include aspects such as a negative research climate, mentoring, or inadequate oversight.

Conclusions

This study, carried out in a Spanish university, shows a significant perception of the proliferation of misconduct, with 71.68% at national level and 48.95% in the institution itself, pointing to similar trends to those seen in other empirical studies. One striking result is that the most commonly perceived misconduct is related to life with colleagues. More specifically, the study highlights the incidence of aspects that have attracted less research, such as the use of personal influence (in assessment or review processes) and the abuse of power over people in lower positions.

The causes behind the proliferation of misconduct are attributed in the closed questions to personal ambitions and pressure from the evaluation system, while in the open questions the issue of the evaluation system is the main focus. This evaluation model, based on the use of bibliometric indicators, is defined as one that encourages ‘publish or perish’, which is having negative consequences for scientific integrity. Thus, the researchers surveyed generally consider it necessary to go to the root of the problem and overhaul the evaluation system and the rationale behind the evaluation and measurement of research merit characteristic of this setting. A range of possible alternatives were suggested to tackle this change, although the calls for a return to qualitative evaluation criteria predominate. A clear indicator of this demand in the Spanish context has also been reflected recently in the COARA initiative.Footnote 7 Spain is one of the countries with the highest number of signatories and with a very prominent presence of universities, which could be explained by the prioritisation of quantitative criteria in the national evaluation system.The data from our study shed light on the motives behind these demands to transform or improve the scientific evaluation system. The measures implemented by ANECA at the end of 2023, involving the approval of new criteria for assessing sexenios (where the significance of bibliometric indicators is diminished) suggest that the review of the assessment system is underway.Footnote 8