Developing student competence through peer assessment: the role of feedback, self-regulation and evaluative judgement
How can students’ competence be developed through peer assessment? This paper focuses on how relevant variables such as participation, evaluative judgement and the quality of the assessment interact and influence peer assessment. From an analysis of 4 years of data from undergraduate classes in project management, it develops a model of causal relationships validated using the PLS-SEM method. It demonstrates relationships between these variables and considerers the influence of students’ competence and the mediating nature of feedback and self-regulation on the process. It points to how peer assessment practices can be improved whilst highlighting how evaluative judgement and feedback are two key elements that can be addressed to deliver the effective development of students’ competence.
KeywordsPeer assessment Feedback Evaluative judgement Self-regulation Structural equation modelling Importance-performance map analysis
In recent years, we have witnessed a progressive evolution of assessment processes that has changed the focus of attention towards students’ strategic and lifelong learning. From the recall of knowledge, emphasis has moved onto students’ ability to respond to assessment tasks that are divergent rather than convergent and complex rather than simple (Sadler 2012). Among the approaches accompanying this refocusing are assessment for learning (Carless et al. 2017; Lai 2006; Sambell et al. 2013), learning-oriented assessment (Carless et al. 2006; Carless 2015), assessment as learning (Dann 2014; Earl 2013), sustainable assessment (Boud and Soler 2016; Nguyen and Walker 2016) and assessment as learning and empowerment (Rodríguez-Gómez and Ibarra-Sáiz 2015).
In these different approaches, particular significance is given to participatory modalities of assessment, such as self-assessment and peer assessment. The reviews carried out by Dochy et al. (1999), Gielen et al. (2011) and Panadero (2016) illustrate the variety of ways in which peer assessment can be implemented in practice.
Given the centrality of peer assessment in reforming assessment practices, the purpose of this paper is to examine how peer assessment practices can be analysed and thereby improved. It develops an exploratory and predictive model that considers the key variables involved in peer assessment. To do this, it discusses theoretical foundations that suggest possible causal relationships between relevant variables. The model developed is validated using data from peer assessment activities with undergraduates in Spain taking an Economics and Business Degree over four academic years. The students experienced peer assessment as part of the course, using the EvalCOMIX® web service, which was expressly developed to promote participatory modes of assessment (Ibarra-Sáiz and Rodríguez-Gómez 2017).
Provide a predictive model of the competence development of students, based on the practice of peer assessment and which illustrates the relationships between variables such as evaluative judgement, participation, feedback, self-regulation and the quality of the assessment.
Offer an instrument that facilitates analysis and understanding of the perception of university students about peer assessment practices using technological resources.
Orientate the practice of peer assessment towards those aspects with the greatest potential for improving students’ competence development.
Framework and development of hypotheses
Practicing peer assessment
Until the 1990s, assessment processes in universities tended to focus on what students knew. Students were assessed, above all, on their understanding of some domain of specific knowledge within the subject area they studied. Progressively, the emphasis has been refocused onto what students can do and the value of transferable, generic or essential skills, that is, the skills and competences that all students should develop (Boud 2014; Strijbos et al. 2015).
As other authors have argued (Nicol et al. 2014; Thomas et al. 2011), learning with peers is assessed because it is a key skill required for lifelong learning, which involves critical thinking and reflection and being able to evaluate one’s own work and that of others. Universities increasingly focus their efforts on these skills so that they represent an essential part of what students learn throughout their university studies.
At the beginning of any peer assessment practice we are faced with a situation in which the role of students needs to substantially change. Giving the student a voice implies modifying the traditional relationship of power in assessment processes, from one in which the lecturer holds a dominant position to a more equitable and democratic relationship in which students assume responsibility themselves as an assessor. Assessment therefore transforms from a unidirectional process, dominated by lecturers, to a socio-constructive and dynamic process in which lecturers and students interact (Rust et al. 2005).
This new situation requires numerous variables to be taken into consideration. Aspects such as participation (Falchikov 2005), students’ evaluative judgement (Boud et al. 2018a, b), self-regulation (Hawe and Dixon 2017; Panadero et al. 2017), feedback (Boud and Molloy 2013; Nicol et al. 2014), a climate of trust (Carless 2009, 2013) and the quality of the assessment (Sadler 2016) are all elements that play a vital role in assessment practice. In this study, we focus primarily on the two elements that can be considered as basic to peer assessment: student participation and their evaluative judgement. Secondly, we analyse the role played by feedback and self-regulation, as well as the value that students attribute to peer assessment in contributing to their competence development.
Participation and evaluative judgement
Contemporary theories acknowledge the central role of the student in the construction of their own learning (Penuel and Shepard 2016). In the case of peer assessment, the importance of student participation has been highlighted by Falchikov (2005), Thomas et al. (2011), Reinholz (2016) and López-Pastor and Sicilia-Camacho (2017). Participation means encouraging dialogue with students and enabling them to collaborate in the process of assessing their learning in ways that are transparent. This participation can be established during all three phases of the assessment process: planning, development and judgement.
During the planning, phase tutors can decide or agree with their students the selection or definition of criteria, the means of assessment, the design of assessment instruments or the grading system. When the assessment is carried out, students can participate by assessing their own work and actions and those of their peers, through assessment modes such as self- assessment or peer assessment. Finally, students can participate fully in the grading process through dialogue and consensus around the grades allocated.
For the students’ judgement on assessed work to be fair and equitable, it is vital that they have significant competence in the practice of assessment. The importance of students’ ability to make judgements has been evidenced by Boud and Falchikov (2007), Cowan (2010), Yucel et al. (2014), Nguyen and Walker (2016) or Murillo and Hidalgo (2017). In fact, Boud (2014, p. 27) highlights the importance of the development of informed judgement as one of the strategic axes in the assessment change agenda, because it is the “‘sine qua non’ of assessment”.
Tai et al. (2018, p. 471) define evaluative judgement as “the ability to make decisions about the quality of work of self and others”. In short, this implies the identification or discernment of standards, the application of them to a given piece of work, techniques for calibrating judgement and mechanisms to avoid being fooled (Boud 2016).
Under this conception of evaluative judgement lies a double dimensionality, in so far as it supposes an assessment of one’s own work and that of others. Furthermore, if we add another determining factor in the assessment process, such as trust or a lack of trust (Carless 2009, 2013), we can consider that, in the context of peer assessment, evaluative judgement is underpinned by trust in one’s own judgement and trust in the judgement of others.
Feedback and self-regulation
In peer assessment, the role of feedback is crucial. In general, research shows that feedback is associated with learning and performance because, as Hounsell (2007) states, feedback can improve learning in three different ways: by accelerating learning, by optimizing the quality of what is learned and by raising the level of achievement of individuals and of groups.
There are numerous contributions to the characteristics of high-quality feedback, as well as suggestions on how further improvements can be made (Ajjawi and Boud 2017, 2018; Boud and Molloy 2013; Espasa and Meneses 2010; Evans 2013; Gielen et al. 2011; Pardo 2018). However, in recent years, we have seen a change in the meaning and purpose of feedback. Previously, special attention was paid to features such as speed, level of detail, clarity, structure or relevance in the delivery of information to students; but nowadays, attention has shifted towards the meaning of feedback for the student and the interaction between the student and the giver of the feedback (Rowe 2017). Feedback has evolved from being perceived as a one-way process of transmitting information from lecturer to student, to being considered as a process using multiple communication channels, through which lecturers and students interact with each other in order to lead to improved outcomes. This highlights the importance of facilitating the participation of students as a source of feedback and learning (Moore and Teather 2013; Nicol et al. 2014)
In the context of peer assessment, it is essential for students to understand what quality feedback involves. They need to learn to evaluate and make judgements about the quality of the work of others whilst maintaining a dialogue with their lecturers and colleagues about the quality of their assessment (Sadler 2012). Consequently, it is important for students to develop the ability to make judgements and evaluate constructively. However, that alone is not sufficient. In addition, students must be able to use the feedback offered in such a way as to reduce the gap between the feedback given and the feedback used (Cartney 2012).
This change of the focus and the actors in feedback implies that lecturers should pay less attention to delivering unidirectional, focused and direct feedback and pay more attention to how students understand and interpret multidirectional feedback from their self-regulatory and self-productive identities (Dann 2014).
The self-regulation of learning is an essential feature of effective learning processes, as supported by multiple existing models (Panadero 2017). One of the assumptions underlying self-regulated learning is the importance of the nature of mediation that takes place between personal and contextual characteristics and the level or degree of achievement or execution (Pintrich 2000; Järvelä et al. 2016). The work of Panadero et al. (2018) shows that many studies have explored the relationships between self-regulated learning, the use of learning strategies and academic performance. However, few studies have focused their attention on the role of assessment as an element conducive to self-regulated learning.
The achievement of high and consistent quality in assessment practices is required by students (Smith and Coombe 2006). With regard to assessment standards, it has been proposed that “Classroom assessment practices meet the standards of quality when teachers can be confident that their assessment practices provide accurate and dependable information about student learning” (Klinger et al. 2015). In fact, the generating of quality evidence is one basic principle in the Berkeley Evaluation and Assessment Research Assessment (BEAR) System, based on four principles: (1) developmental perspective, (2) a match between instruction and assessment, (3) the generating of quality evidence, and (4) management by instructors to allow appropriate feedback and follow-up (Wilson and Scalise 2006, p. 646). However, this is a partial view since it only takes into consideration the perspective of teachers. Quality assessment should be perceived by students as a rigorous evaluation, that is both valuable and interesting. Assessment quality also relates to other aspects mentioned previously, such as trust in the judgement of others and the usefulness of their judgements.
During peer assessment, if students do not receive pertinent, constructive feedback from their peers, there is a risk that they will perceive their judgements as unfair and discouraging. It is vitally important, therefore, that quality feedback information is given during peer assessment to reduce any perception of injustice and increase students’ motivation and commitment (Moore and Teather 2013).
Research model and hypotheses
Competences are an integration of knowledge, skills and attitudes and are situated in or aligned to a professional context
Specific and detailed information on the quality of work and improvement in commitment aimed at enabling improvement
Self-generated thoughts, feelings and actions planned and cyclically adapted to the attainment of personal goals
Quality of assessment
Assessment that is rigorous, credible, interesting and useful for learning
The ability to make decisions about the quality of work of self and others
Design and wording of assessment criteria and instruments to assess their own work and that of their peers
H1. Evaluative judgement consists of two components; trust in one’s own judgement and confidence in the judgement of others
H2a. Evaluative judgement is expected to be positively related to the quality of the assessment
H2b. Participation is expected to be positively related to the quality of the assessment
H3a. Participation is expected to be positively related to the development of competence
H3b. Evaluative judgement is expected to be directly related to the development of competence
H3c. Self-regulation is expected to be positively related to the development of competence
H3d. Feedback is expected to be positively related to the development of competence
H4a. The relationship between evaluative judgement and the development of competence is expected to be mediated by feedback.
H4b. The relationship between participation and the development of competence is expected to be mediated by feedback
H5a. The relationship between evaluative judgement and the development of competence is expected to be mediated by self-regulation
H5b. The relationship between participation and the development of competence is expected to be mediated by self-regulation
H6a. The relationship between participation and the development of competence is expected to be mediated by the quality of the assessment
H6b. The relationship between evaluative judgement and the development of competence is expected to be mediated by the quality of the assessment
This study was carried out during four academic years (2012/2013, 2013/2014, 2016/2017 and 2017/2018) using peer assessment as an integral part of the assessment process in the subject Project Management. A cohort design for the research was used, since the self-perception of the students was collected at the end of each semester. Consequently, different students responded in each academic year. A set of three assessment tasks was designed in which the students had to perform both self-assessment and peer assessment of the products or actions that were being assessed.
Peer assessment in practice
Assessment tools used for every assessment task
Design and completion of a comparative report on documents relating to Business and Economics
• Rubric for comparative reports on professional documents
• Rating scale for oral presentations
Oral presentation on predictive and agile methodologies for Project Management
• Rating scale for oral presentations on methodologies
Design and planning of a project in the area of Business and Economic Science
• Rating scale for project assessment
The EvalCOMIX® web service (Ibarra-Sáiz and Rodríguez-Gómez 2017), integrated into the Moodle server of the university’s virtual campus, was used to design, manage and apply all assessment instruments. Using EvalCOMIX® facilitated the peer assessment process, the delivery of feedback and the final calculation of grades, based on the criteria and weighting of each of the elements within the assessment task.
In order to obtain the students’ views on participating in assessment processes through peer assessment, we designed an ad hoc questionnaire, the “Student Perception of Peer Assessment in Practice” questionnaire (Appendix D) (Online Resource 4).
Participation satisfaction questionnaire structure
No. of items
Quality of the assessment
Confidence in their own judgement
Confidence in the judgement of others
Partial least squares structural equation modelling (PLS-SEM) method and the statistical software SmartPLS 3 (Ringle et al. 2015) were used. This is a second-generation technique designed to overcome the weakness in more traditional exploratory type first-generation methods such as cluster analysis, exploratory factor analysis or multidimensional scaling (Hair et al. 2017). PLS-SEM is used specifically to develop theories in exploratory research by focusing on the explanation of variances in dependent variables when analysing a model.
PLS-SEM is recommended when, as here, the objective is the prediction of an objective construct or the intention is to identify “driver constructs”; the research model is complex according to the type of relationships hypothesised (direct and mediation) and the levels of dimensionality (first-order and second-order constructs); formatively measured constructs are part of the structural model; the structural model is complex and the data is non-normal (Hair et al. 2017).
Confirmatory tetrad analysis (CTA-PLS) was employed to confirm the formative or reflective nature of the constructs. This is used to check the adequacy of the specification of the measurement model and test the null hypothesis that the indicators for a model are reflective (Garson 2016), so that the reflective or formative nature of the latent variables can be confirmed (Hair et al. 2018).
The evaluation of the model has been carried out according to the reflective (Mode A) or formative (Mode B) character of the model. Once the measurements of the constructs were confirmed as reliable and valid, we proceeded to analyse the predictive capacity of the model and the relationships between the constructs.
Finally, the importance-performance map analysis (IPMA) technique was used to identify predecessor constructs that have a relatively high importance for predicting the target construct, but “also have a relatively low performance so that improvements can be implemented” (Hair et al. 2018, p. 105). This technique allows constructs and indicators to be easily identified so they can be modified to improve results in an effective way and enable improvements or changes to be prioritised.
Descriptive statistics and contrast tests (Mann-Whitney U and Kruskal-Wallis)
Competency development (CODEVP)
Quality of the assessment (QUASSP)
Confidence in their own judgement (OWNJUP)
Confidence in the judgement of others (OTHJUP)
We initially considered all the constructs to be formative but, after carrying out confirmatory tetrad analysis (CTA-PLS), no empirical evidence of this could be seen for the constructs feedback (FEEDFP), confidence in their own judgement (OWNJUP) and confidence in the judgement of others (OTHJP) so the decision was taken to consider them as reflective.
Evaluation of the measurement model
(a) Reflective model
The evaluation of the measurement model for reflective indicators in PLS-SEM is based on internal consistency reliability, convergent validity and discriminant validity (Hair et al. 2017). As the values of internal consistency reliability and Cronbach’s Alpha values are above the 0.70 threshold, we can conclude that the four constructs are reliable. Average variance extracted (AVE) values for latent variables are greater than 0.61. Thus, the measure of the four reflective constructs has high levels of convergent validity. The Heterotrait-Monotrait Ratio (HTMT) uses 0.85 as the relevant threshold level, a criterion which is also met in this study. This means that all the constructs are empirically distinct (Online Resource 5).
(b) Formative model
With values of variance inflation factor (VIF) between 1.43 and 4.06, we can conclude that collinearity does not reach critical levels in any of the formative constructs and is not an issue for the estimation of the PLS path model (threshold value of 5). Some indicators were found whose values were not statistically significant but instead had loads greater than 0.5, so according to the rules of thumb expressed by Hair et al. (2017, p. 151), all the formative indicators were maintained (Online Resource 6).
Evaluation of structural model
Structural model results using t values and percentile bootstrap 95% confidence interval (n = 5000 subsamples)
Path coefficients (*)
Effect size (+)
JUDGEP → OTHJUP
JUDGEP → OWNJUP
JUDGEP → QUASSP
PARTI → QUASSP
PARTI → CODEVP
[− 0.065, 0.031]
JUDGEP → CODEVP
[− 0.063, 0.039]
SELFRP → CODEVP
[− 0.026, 0.089]
FEEDF → CODEVP
We can confirm the predictive value of the model through the analysis of the coefficient of determination (R2). Thus, it is evident how 65.9% of the variance (R2) of the competence development construct (CODEVP) is explained by four essential constructs. The strongest effect is exercised by the feedback construct (FEEDFP, 0.429), followed by self-regulation (SELFRP, 0.261), evaluative judgement (JUDGEP, 0.140) and participation (PARTI, 0.128).
Evaluative judgement (JUDGEP) is a hierarchical component model (HCM) constructed under a repeated indicators approach (Hair et al. 2017, 2018). That is to say, it is a multidimensional construct, formed by the confidence in one’s own judgement (OWNJUP) and trust in the judgement of others (OTHJUP). Our research model achieves SRMR of 0.072 (Fig. 2), which means an appropriate fit, taking into account the usual cut-off of 0.08.
All Stone-Geisser’s Q2 values (predictive relevance) for endogenous constructs are considerably above zero (Online Resource 7). More precisely, evaluative judgement (JUDGEP) has the highest Q2 value (0.530), followed by feedback (FEEDF, 0.397), quality of the assessment (QUASSP, 0.404), competency development (CODEVP, 0.350) and, finally, self-regulation (SELFRP, 0.347). These results provide clear support for the model’s predictive relevance regarding the endogenous latent variables. A medium value is reached by effect sizes (q2) in the cases of evaluative judgement (JUDGED) on quality of the assessment (QUASSP) and participation (PARTI) on feedback (FEEDFP), with smaller values in the other cases.
(a) Feedback as mediator
Summary of mediating effect test of PARTI on CODEVP
PARTI → CODEVP
PARTI → CODEVP
Total indirect effect
PARTI → CODEVP
Specific indirect effect
PARTI → FEEDF → CODEVP
PARTI → QUASSP → FEEDF → CODEVP
PARTI → FEEDF → SELFRP → CODEVP
PARTI → QUASSP → FEEDF → SELFRP → CODEVP
PARTI → SELFRP → CODEVP
[− 0.003, 0.072]
PARTI → QUASSP → SELFRP → CODEVP
Summary of mediating effect test of JUDGED on CODEVP
JUDGEP → CODEVP
JUDGEP → CODEVP
Total indirect effect
JUDGEP → CODEVP
Specific indirect effect
JUDGEP → FEEDF → CODEVP
JUDGEP → QUASSP → FEEDF → CODEVP
JUDGEP → FEEDF → SELFRP → CODEVP
JUDGEP → QUASSP → FEEDF → SELFRP → CODEVP
JUDGEP → SELFRP → CODEVP
[− 0.037, 0.050]
JUDGEP → QUASSP → SELFRP → CODEVP
So far, we have presented the results of the analysis of simple mediation, but as we can see in Fig. 2 and Tables 7 and 8, feedback also operates in the context of multiple mediation, that is, in the mediation that occurs when an exogenous construct exerts its influence through more than one mediating variable. This multiple analysis allows us to consider all the mediators simultaneously in one model, enabling us to obtain a better representation of the mechanisms through which an exogenous construct affects an endogenous construct (Hair et al. 2017).
In the case of our model, we can see in Table 7 how feedback also intervenes in this multiple mediation, between participation and competence development, together with the quality of the assessment (0.049) and self-regulation (0.024). This multiple mediation of feedback between evaluative judgement and competence development is also evident (Table 7), together with the quality of the assessment (0.111) and self-regulation (0.022 and 0.024). In all cases, it is a partial (complementary) mediation.
(b) Self-regulation as mediator
Regarding the mediation of self-regulation between participation and competence development (Table 7), we find that the value of the specific indirect effect is not significant (0.027). Since the direct effect is significant and this indirect effect is not significant, we conclude that there is no mediation, but when analysing multiple mediation, the results indicate that it occurs in conjunction with the quality of the assessment (0.027), with feedback (0.024) and with the quality of the assessment and the feedback together (0.011). In this case, it would be a partial (complementary) mediation.
In the case of the mediation of self-regulation between evaluative judgement and competence development (Table 8) something similar happens. The results indicate that there is no simple mediation (0.004). On the other hand, it is evident that it mediates in conjunction with feedback (0.022), the quality of the assessment (0.061) and with both together (0.024). In this case, there is also a partial (complementary) mediation. Therefore, within the context of collaboration, the process by which students critically analyse their own work and that of their peers and identify omissions or errors that help them improve their own and their peers’ work is seen to be a valuable element in developing their competences.
Importance-performance map analysis
In addition to the evaluation of the measurement model, the structural model and the analysis of simple and multiple mediation, an importance-performance map analysis (IPMA) was carried out (Online Resource 8). The rationale of IMPA is “to identify predecessor constructs that have a relatively high importance for predicting the target construct, but also have a relatively low performance so that improvements can be implemented” (Hair et al. 2018, p. 105). In our case, the constructs on which action could be taken to improve competence development are, firstly, feedback and evaluative judgement, followed by participation. If we make self-regulation the objective, we could act on the quality of the assessment. And, finally, if we take as an objective the improvement of feedback, the variable which we should act on is evaluative judgement.
In this paper, we intended, firstly, to provide a predictive model of students' competence development based on the practice of peer assessment. Secondly, we sought to propose an instrument through which to analyse and understand the perception of university students about peer assessment practices using technological resources. Finally, we wanted to guide the practice of peer assessment towards formats with the greatest potential for change and improvement. The results achieved in this study suggest there are important implications, both from a theoretical and practical perspective, to understanding peer assessment processes. At the same time, they also provide insight into future lines of research.
The primary objective of this study was to provide a predictive model of students’ competence development based on the practice of peer assessment. In this regard, our study reflects the proposals made by Panadero et al. (2018) who suggest that an analysis is needed of the influence that different models of formative assessment and self-regulation have on each other, what practices considered as formative can promote self-regulated learning and under what conditions.
One of the main contributions of this work is the construction of a model that integrates the relationships between significant variables of peer assessment in a university context. The results obtained demonstrate that the hypothesised model can, indeed, predict a large part of the relationships between the variables and show, on the one hand, that participation and evaluative judgement are directly related to competence development and, on the other, the mediating role of feedback and self-regulation in the context of peer assessment.
Hypothesis H1, in which the construction of evaluative judgement is achieved through two constructs: trust in one’s own judgement and trust in the judgement of others, has been tested and verified. Likewise, hypothesis H2, that directly relates evaluative judgement and participation with the quality of the assessment, as well as hypothesis H3 regarding the positive relationship between participation, evaluative judgement, self-regulation and feedback on the one hand with competence development on the other, have also been confirmed. Finally, the remaining hypotheses concerning the mediating character of feedback (H4), self-regulation (H5) and quality of evaluation (H6) have also been established.
One of the essential purposes of assessment as learning is that students should become the protagonists of their learning (Coombs et al. 2018; DeLuca et al. 2016). This means tutors must assume more of a role as facilitator. Our study confirms student participation in assessment as a variable that is directly related to their competence development and which exerts a direct influence on other aspects such as feedback or self-regulation.
It has been shown in this study how evaluative judgement, in terms of trust in one’s own judgement and in the judgement of others, is directly related to students’ competence development, as well as to feedback and self-regulation. The systematic development of evaluative judgement is currently an important challenge for the university curriculum, since it places it at the very centre of education (Boud et al. 2018b).
A second objective that has guided our research has been to develop a technology-enabled instrument to analyse and understand the perception of university students about peer assessment practices. In relation to this, our evaluation of the measurement model employed supports the validity of the questionnaire used to operationalise the latent variables, since the items are relevant and all the items load on the right construct. As a result, tutors now have access to an easy-to-use instrument through which they can collect students’ perceptions on the implementation of peer assessment.
In this research, the peer assessment process has been carried out using the EvalCOMIX® web service, which allows for greater speed and efficiency in the assessment process. It requires tutors to design and manage the assessment instruments used and monitor the process to address any problems students have with it. This web service can be an excellent technological tool to facilitate the impetus for change identified by Bearman et al. (2017)
The last objective that guided this study was to influence the practice of peer assessment to focus on areas of greatest potential for change and improvement. The results obtained in the IPMA analysis confirm the importance of evaluative judgement and feedback as the primary elements on which to act in order to significantly improve competence development. These results are consistent with the contributions of authors such as Boud et al. (2018a), Dawson et al. (2018), Hernández (2012), Nicol et al. (2014), Rodríguez-Gómez and Ibarra-Sáiz (2015) and Sadler (2016). Consequently, the importance of the mediating role of tutors is crucial. They must foster a climate of trust among students that allows them to carry out a rigorous, credible and objective assessment, whilst providing useful and relevant information for the improvement of future activity. This requires lecturers to educate students about assessment so that they, in turn, can participate and deliver judgements that can be reviewed and contrasted, allowing them to progressively acquire greater confidence in their own judgements and those of their peers.
Limitations and future research
From a methodological perspective, the research described in this paper suffers from three specific limitations. In the first place, the research was carried out in a specific context, with final year project management students in Spain. Research needs to be undertaken in other subject areas and with students at other stages of their studies. Secondly, it is an investigation carried out on the basis of a design with a post-test measurement, meaning the degree of control over the intervening variables is reduced and, in line with the caution advised by Stone-Romero and Rosopa (2008), the inferences that can be made about the mediation model are limited. Finally, the measurement instrument is based on the perception of the students themselves, which, as indicated by Panadero et al. (2018), could be improved by the use of alternative measuring instruments.
To be able to generalise our results more widely, further studies need to be carried out using experimental designs in which both the independent variable (the practice of peer assessment) and the mediating variables (essentially feedback and self-regulation) can be manipulated. Despite these limitations, though, we have been able to verify the great diversity and variability of current assessment practices. This diversity and variability make comparison and generalisation difficult, especially in the context of formal education where experimental studies are often difficult to carry out, but the modes of self-assessment and peer assessment in particular seem to be an issue in the future (Pereira et al. 2017).
Struyven et al. (2005) highlight that students’ perceptions serve to guide us in our reflective attempts to improve our educational practices and achieve a higher quality of learning and education for our students. However, a second line of research would be to improve the instrumentation used in the data collection process, incorporating other ways of collecting the students’ perceptions and collecting data from sources other than students, whilst incorporating measurements that combine both measurement and intervention (Panadero et al. 2016).
The variables involved in the process of peer assessment are highly complex and interact in so many different ways that it is important to try and achieve a greater level of detail, precision and understanding of them. It would therefore be valuable if further research was undertaken using mixed methodologies which, on the basis of explanatory sequential design, could explain the quantitative results in terms of qualitative data obtained. As Creswell and Clark (2010, p. 82) argue, this type of design “is most useful when the researcher wants to assess trends and relationships with quantitative data but also be able to explain the mechanism or reasons behind the resultant trends”.
In this paper, we have demonstrated how the practice of peer assessment is perceived by students as an element that promotes their competence development. We have devised tools that can facilitate adaptation or replication in other different contexts and have suggested future lines of research that will lead to further improvements in assessment. Likewise, we have shown how the implementation of participatory assessment involves a series of interrelationships between different aspects, highlighting the need to address the improvement of feedback processes and the development of evaluative judgement. In fact, providing a context where assessment processes are rigorous, credible, objective and participative, as well as delivering useful and relevant information for the ongoing development of peer assessment practices.
It is vital that policies are developed in higher education that encourage the creation of contexts in which peer assessment can be incorporated, both from a pedagogical and conceptual perspective and also from a technological perspective. Lecturer education on these practices should be promoted and technological resources provided so that the implementation of participatory assessment methods does not become a continuous struggle to overcome bureaucratic difficulties or technological limitations, which can often frustrate and limit educational improvement and change.
This work was supported by the DevalS Project funded by the Ministry of Economy and Competitiveness (Ref: EDU2012-31804) and the TransEval Project (Ref: R + D + i 2017/01) funded by University of Cadiz.
- Boud, D. (2014). Shifting views of assessment: from secret teachers’ business to sustaining learning. In C. Kreber, C. Anderson, N. Entwistle, & J. McArthut (Eds.), Advances and inovations in university assessment and feedback (pp. 13–31). Edinburgh: Edinburgh University Press Ltd.. https://doi.org/10.3366/edinburgh/9780748694549.003.0002.CrossRefGoogle Scholar
- Boud, D. (2016). Current influences on changing assessment: implications for research to make a difference. In EARLI SIG1 Conference. Munchen.Google Scholar
- Boud, D., & Molloy, E. (Eds.). (2013). Feedback in higher and professional education. London: Routledge.Google Scholar
- Boud, D., Ajjawi, R., Dawson, P., & Tai, J. (Eds.). (2018a). Developing evaluative judgement in higher education. Assessment for knowing and producing quality work. London: Routledge.Google Scholar
- Boud, D., Dawson, P., Tai, J., & Ajjawi, R. (2018b). Creating an agenda for developing students’ evaluative judgement. In D. Boud, R. Ajjawi, P. Dawson, & J. Tai (Eds.), Developing evaluative judgement in higher education: assessment for knowing and producing quality work (pp. 186–195). London: Routledge.CrossRefGoogle Scholar
- Carless, D. (2013). Trust and its role in facilitating dialogic feedback. In D. Boud & E. Molloy (Eds.), Feedback in higher and professional education (pp. 90–103). London: Routledge.Google Scholar
- Carless, D., Bridges, S. M., Chan, C. K. Y., & Glofcheski, R. (Eds.). (2017). Scaling up assessment for learning in higher education. Singapore: Springer.Google Scholar
- Cartney, P. (2012). Exploring the use of peer assessment as a vehicle for closing the gap between feedback and feedback used. In S. Hatzipanagos & R. Rocho (Eds.), Approaches to assessment that enhance learning in higher education (pp. 61–72). London: Routledge.Google Scholar
- Creswell, J. W., & Clark, V. P. (2010). Designing and conducting mixed methods research (2nd ed.). Thousands Oaks: SAGE Publications.Google Scholar
- Earl, L. (2013). Assessment as learning. Using classroom assessment to maximize student learning. London: Sage.Google Scholar
- Falchikov, N. (2005). Improving assessment through student involvement. Practical solutions for aiding learning in higher education and further education. London: RoutledgeFalmer.Google Scholar
- Garson, G. D. (2016). Partial least squares: regression & structural equation models. Asheboro: Statistical Publishing Associates.Google Scholar
- Hair, J. F., Hult, G. T. M., Ringle, C. M., & Sarstedt, M. (2017). A primer on partial least squares structural equation modeling (PLS_SEM). London: Sage.Google Scholar
- Hair, J. F., Sarstedt, M., Ringle, C. M., & Gudergan, S. P. (2018). Advanced issues in partial least squares structural equation modeling. London: Sage.Google Scholar
- Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research, 77(1), 81–112.Google Scholar
- Hounsell, D. (2007). Towards more sustainable feedback to students. In D. Boud & N. Falchikov (Eds.), Rethinking assessment in higher education: Learning for the longer term (pp. 101–113). London: Routledge.Google Scholar
- Ibarra-Sáiz, M. S., & Rodríguez-Gómez, G. (2017). EvalCOMIX®: a web-based programme to support collaboration in assessment. In T. Issa, P. Kommers, T. Issa, P. Isaías, & T. B. Issa (Eds.), Smart technology applications in business environments (pp. 249–275). Hershey: IGI Global. https://doi.org/10.4018/978-1-5225-2492-2.ch012.CrossRefGoogle Scholar
- Järvelä, S., Järvenoja, H., Malmberg, J., Isohätälä, J., & Sobocinski, M. (2016). How do types of interaction and phases of self-regulated learning set a stage for collaborative engagement? Learning and Instruction, 43, 39–51. https://doi.org/10.1016/j.learninstruc.2016.01.005.CrossRefGoogle Scholar
- Johnson, R. L., & Morgan, G. B. (2016). Survey scales. A guide to development, analysis, and reporting. London: The Guilford Press.Google Scholar
- Klinger, D., McDivitt, P., Howard, B., Rogers, T., Muñoz, M., & Wylie, C. (2015). Classroom assessment standards: practices for PK-12 teachers. Seattle: Kindle Direct Press.Google Scholar
- Moore, C., & Teather, S. (2013). Engaging students in peer review : feedback as learning. Issues in Educational Research, 23(2), 196–212.Google Scholar
- Panadero, E. (2016). Is it safe? Social, interpersonal, and human effects of peer assessment: a review and future directions. In G. T. L. Brown & L. R. Harris (Eds.), Handbook of social and human conditions in assessment (pp. 247–266). New York: Routledge.Google Scholar
- Penuel, W. R., & Shepard, L. A. (2016). Social models of learning and assessment. In A. A. Rupp & J. P. Leighton (Eds.), Handbook of cognition and assessment: frameworks, methodologies, ans applications. Hoboke: Wiley.Google Scholar
- Ringle, C. M., Wende, S., & Becker, J.-M. (2015). SmartPLS 3. Bönningstedt: SmartPLS. http://www.smartpls.com.
- Rodríguez-Gómez, G., & Ibarra-Sáiz, M. S. (2015). Assessment as learning and empowerment: towards sustainable learning in higher education. In M. Peris-Ortiz & J. M. Merigó Lindahl (Eds.), Sustainable learning in higher education. Developing competencies for the global marketplace (pp. 1–20). Springer International Publishing. https://doi.org/10.1007/978-3-319-10804-9_1.Google Scholar
- Rowe, A. D. (2017). Feelings about feedback: the role of emotions in assessment for learning. In D. Carless, S. M. Bridges, C. K. Y. Chan, & R. Glofcheski (Eds.), Scaling up assessment for learning in Higher Education (pp. 159–172). Singapore: Springer. https://doi.org/10.1007/978-981-10-3045-1_11 Google Scholar
- Sadler, D. R. (2012). Beyond feedback: developing student capability in complex appraisal. In S. Hatzipanagos & R. Rochon (Eds.), Approaches to assessment that enhance learning in higher education (pp. 45–60). London: Routledge.Google Scholar
- Sambell, K., McDowell, L., & Montgomery, C. (2013). Assessment for learning in higher education. London: Routledge.Google Scholar
- Skjong, R., & Wentworth, B. (2000). Expert judgement and risk perception. Hovik: Det Norsrke Veritas. http://research.dnv.com/skj/Papers/SkjWen.pdf. Accessed on 02/09/2018.
- Tai, J. H. M., Canny, B. J., Haines, T. P., & Molloy, E. K. (2016). The role of peer-assisted learning in building evaluative judgement: opportunities in clinical medical education. Advances in Health Sciences Education, 21(3), 659–676. https://doi.org/10.1007/s10459-015-9659-0.CrossRefGoogle Scholar
- Thomas, G., Martin, D., & Pleasants, K. (2011). Using self-and peer-assessment to enhance students’ future-learning in higher education. Journal of University Teaching & Learning Practice, 8(81).Google Scholar
- Yucel, R., Bird, F. L., Young, J., & Blanksby, T. (2014). The road to self-assessment: exemplar marking before peer review develops first-year students’ capacity to judge the quality of a scientific report. Assessment & Evaluation in Higher Education, 39(8), 971–986. https://doi.org/10.1080/02602938.2014.880400.CrossRefGoogle Scholar
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.