Background

Among many factors influencing student learning, teacher quality is the most determining one (Snook et al. 2013). Teaching quality is an important criterion for quality assessment of education utilized by students, parents, and authorities (Feistauer and Richter 2016) as it is the most determining factor in students’ achievement (Sanders et al. 1997). Research indicates that teachers improve their teaching quality by acting on students and authority evaluations (Dresel and Rindermann 2011). There is little attention on formative teacher assessment, and the existing studies on teacher evaluation are either certificate or accomplishment and product-based evaluation based on students’ scoring. The problem with this type of evaluation is that they do not provide any information about the teaching practice teachers do (Bastian et al. 2016; Henry et al. 2010). Although new approaches have been introduced to the field of teacher education and evaluation, it has been a long time that teacher evaluation was through the students’ assessment of teachers’ teaching. This evaluation has been conducted through teacher evaluation questionnaires (Marsh et al. 2009) which are under question for reliability concerns (Feistauer and Richter 2016). Although later classroom observation, student evaluation questionnaire, teacher individual interviews, teacher self-evaluation, and teacher testing (Santiago and Benavides 2009; Smith et al. 2004) were introduced to the field of teacher evaluation, they provide little insights about how to improve teaching practice (Duckor et al. 2014). The inadequacy of the product-based approaches towards teacher evaluation led practitioners to think of inventories that focus more on process-based teacher evaluation and real teaching act (Navidinia et al. 2015). This study is intended to design an inventory for evaluating teacher competences and its potential in tracking changes in the actual teaching act.

Background and purposes

Teacher competences

Teacher development is defined as teachers’ construction of teaching competences (Avalos 2011). Competence is defined as a set of professional skills that underlie successful performances (Blašková et al. 2014). Avalos (2011) stated that teacher competence is the teachers’ ability in critical analysis of teaching phenomena and education policies which enables them to design the teaching process and procedure in a way to achieve the objectives. Duţă et al. (2014) also state that competence is the ability to use skills and knowledge in a coherent and dynamic way to solve problems efficiently. Accordingly, competence is defined by three dimensions: cognitive dimension (knowledge), functional dimension (skills), and attitudes and value dimension (teacher autonomy and responsibility).

Zimpher and Howey (1987) describe four teacher competences: “(1) clinical competence (practical reading and problem solving), (2) personal competence (understanding of self from multiple perspectives with expertise in interactive capacities in interpersonal interactions), (3) critical competence (disposition to engage in social critique and reconstruction of repressive practices), and (4) technical competence (determining in advance what is to be learned and how it is to be learned and criteria by which success is to be measured”.(p. 103). There is a great deal of diversity in terms of what performances are indicators of competences. Lasauskienė et al.’s (2015) action research verified teaching practices and performances that relied on teacher competences. Competency Framework for Teachers proposed by the Department of Education and Training in Australia (2004) is another project in finding the indicators of competences in performance. It has been claimed that although teacher evaluation has received special attention around the world, teachers have been provided with least support for self-evaluation since educators are unaware of the potential evaluation and support tools (Alamoudi and Troudi 2017). The inefficacy of student evaluation of teachers through questionnaires and alternatives such as one-dimensional classroom observation of teaching practice, teacher interviews and self- assessment and teacher portfolio writing were rather product-based evaluations rather than a process-based one (Imhof and Picard 2009) since it was indicated teachers’ focus on immediate performance rather than understanding underlying processes (Mansvelder-Longayroux et al. 2007). Wei’s (2015) study of formative (classroom observation) and summative (student survey of teaching quality) also indicated that when there is no clear feedback on and definition of what good teaching practice is, summative and formative assessment are meaningless and less effective for teachers, students, and high stakes.

Developing a measure for teacher competences

Many ways are suggested about how to measure and help teachers further develop their teaching competences. Most of the teacher evaluation programs were based on students’ achievement scores, and they provide no information about specific teaching practices teachers do, no information for teachers to identify the problems stem from programs, and provide no evidence for teacher performances (Bastian et al. 2016; Henry et al. 2010). Although recently, other measurement instruments such as classroom observations and questionnaires are advocated by the education researchers (Henry et al. 2010), they all suffer from a problem; they come too late to help teachers improve (Bastian et al. 2016). Student rating process does not show goal attainment, increase teacher effectiveness, and student learning (Hughes and Pate 2012). The validity of using non-academic measures as students rating is under question since studies find a positive relation between students’ scores and their rating to the teachers.

Teacher performance evaluation at high stakes measures either for decisions on certification or program completion and adaptations (Duckor et al. 2014). Bastian et al. (2016) compared locally and officially scored performance assessment, and the results of their study indicated that local scores were higher than official scores. However, to make high stake decisions, locally scored performance assessment is not appropriate. It is more logical to have both local and official scoring performance assessments; local scoring performance assessment can provide language, context, and evidence-based evaluation, and official scoring evaluation performance can provide information about if it has construct validity, predictive validity, and reliability.

In a study, Moreno-Murcia et al. (2015) designed and validated a measuring instrument to evaluate the performance of university students, and through factor analysis, they have found that there are three important performances that are considered to be important: (1) planning which refers to previous reflection and designing of the teaching including planning of courses, learning activities, and evaluation criteria; (2) development of the course which is anything related to execution of and compliance with education curriculum; (3) results which refer to the achievement of objectives, achievements of the students, revisions and improvement of teaching activities, and creation of teaching materials.

The inadequacy of product-based and certification approaches led the practitioners to use a more process-oriented evaluation (Imhof and Picard 2009). Among process-oriented evaluation techniques, portfolio assessment has received good attention. In their study of what makes a portfolio and its effect determining in teacher education classes, they assert that portfolio assessment should be an integral part of the education environment and be valued by supervisors and teachers, and they should be given feedback; otherwise, they will consider that portfolios are tedious, time-consuming, and ineffective.

Admiraal et al. (2011) used video portfolios to assess teacher performances and analyzed the reliability and construct and consequential validity of this instrument. They highlight the qualitative and contextual information they provide for the researchers. In their study of reliability and validity concerns attributed to video portfolio instrument for assessing teacher competences, it was established that although there were problems considering reliability and validity of video portfolio as instrument for data collection, teacher assessors rated them positively, and several techniques of think-aloud sessions and reflection session helped the researchers cater for reliability and validity issues. E-portfolio also indicated an increase in teacher reflection and collaboration (Hooker 2017). Pre- and post-interview, reflective journals, and recoding of professional learning community intervention indicated that experienced teachers’ self-efficacy was improved in terms of more use of innovative teaching strategies and language proficiency, and novice teachers were improved in terms of classroom management and autonomy (Zonoubi et al. 2017).

A more recent research conducted by Hughes and Pate (2012) suggests a teacher balanced scorecard as an instrument to evaluate teacher education induction programs. They stated that balanced scorecard is mostly used by organizations to manage their customer services by “translating the organization’s strategy and vision to objectives and measures and targets from finical, customer, internal business processes perspectives” (p. 59). They worked on the possibility of changing classic balanced scorecard into teaching balanced scorecard. Table 1 shows what information can teacher balanced scorecard provide form different perspectives including institutional, departmental/administrative, and learning and growth perspectives—the perspectives introduced by Hughes and Pate (2012).

Table 1 The classic balanced scorecard (BSC) versus the teaching balanced scorecard (TBSC)

As it is indicated in Table 1, teaching balanced scorecard (TBSC) is a multiple measures of teacher education from various perspectives. It is a talking paper that helps the teachers and faculty to communicate and for the faculty to convey expectation it has from the teachers and addresses those aspects of teaching that are beyond the students’ capacity to rate. The classic balanced scorecards are developed mostly for measuring the adequacy of the functioning of organizations from managerial perspectives for the purpose of maximizing product sells and higher income. Therefore, there is a need for a modified balanced scorecard to be used in teacher education agenda tapping how teacher competences and, in turn, teacher performances can be improved through inductions. This study aimed at preparing a teacher balanced scorecard and assessing its reliability and construct validity as a potential instrument for teacher evaluation.

Methods

Participants

Teachers

A randomly selected sample of 216 Iranian EFL male (n = 98) and female (n = 118) teachers’ teaching acts was evaluated by three supervisors. Teachers had more or less the same years of teaching experience (m = 5) and they ranged 26–32 in age (m = 29). All teachers were MA graduates in EFL. They were duty-paid job English teachers in language College of the Researcher’s institution. They were required by the institution to follow the same educational objectives through the same educational materials. This research was a self-funded project. To observe the ethics in research, teachers were informed about the research and were assured that their responses were confidential and would only be used for research purposes, and they signed a consent form for the perusal of their responses in this project. The research deputy of the researcher’s institution (Dr. Reza Ezati—the deputy of research and technology) can approve actions on ethical consideration in this research project.

Supervisors

An invitation letter was sent to the three supervisors from the three institutes. They had the same years of supervisory experience (m = 7) in teaching English as a foreign language (TEFL) centers. They were also Ph.D. holders in TEFL. Since the supervisors were the students of leading researcher and there was the risk of their compulsory participation in research because of power relation and respects they had towards her, they were assured that their decline to participate would not affect their relationship. Therefore, the supervisors’ voluntary participation would assure their motivation and serious endeavor and effort they put into action. They evaluated teachers on TBSC through portfolio writing. The inter-rater reliability Cronbach alpha level of 0.78 indicted reliability of decisions made on TBSC assessment and portfolio writing.

Panel of experts

Five Iranian male (n = 1) and female (n = 4) assistant professors in TEFL from researchers’ institution made the panel of experts. They contributed to the study at two phases: (a) designing the themes and indicators of teacher competences and (b) arranging the competence indicators in teacher balanced scorecard (TBSC).

Instrument

To investigate whether the teacher inventory was effective in detecting teacher competences in tracking competence developments, the researcher asked the teachers to write teacher portfolios in three occasions of the beginning, middle, and end of the semester. Teacher writing portfolio consisted of reflective evaluation of their growth, references to the evidences of growth by providing the best exemplar from the archive of teaching they have, their future vision of the problems they have in teaching and how they are going to solve them, and their evaluation of feedback they received from the mentors and how they respond to the comments. Portfolio writing had checkpoints for teachers, reflection prompts by which teachers’ reflection is directed to have an appropriate account of their progress, and an area for reviewing portfolios and checking grades of portfolio assessment. Several suggestions on how to interpret the themes and how to provide requested information were provided for each theme.

Procedure in data collection and analysis

To compile items of the TBSC inventory, the panel of experts reviewed the literature on teacher competences. Four teacher competences were identified. The definition of each competence was carefully studied to identify unique characteristics of each competence. Four teacher competences including clinical, technical, personal, and critical from three perspectives of student, departmental/administrative, learning and growth were identified. The indicators of competence in practice were derived from the literature and operationalized through Competency Framework for Teachers proposed by the Department of Education and Training in Australia (2004). In operationalizing the indicators of teacher competences, a focused group discussion was conducted by the panel of experts to assess the appropriateness of each indicator of four teacher competences not only with respect to its transparency and relevance but also in terms of the appropriateness of locating them in the right perspective measures. The deigned TBSC had 65 items rated on five Likert scales of unacceptable, slightly unacceptable, neutral, slightly acceptable, and acceptable points (Additional file 1). Table 2 displays the structure of the TBSC questionnaire.

Table 2 Structure of teachers balanced scorecard

Three supervisors examined the TBSC inventory with 216 teachers. Their evaluation of the teachers’ portfolio writing on three occasions (beginning, middle, and the end of the semester).The portfolio of teachers was assessed using Bakker et al.’s (2011) schemata which required supervisors to look for negative and positive evidences of teacher competence, look for (counter) evidences of what contributes to professional thinking and acting, differentiate less and more important evidences and assign score, specify if entire performance can be attributed to specific level of competence, and write a brief summary in which comments on scores were given and important arguments and evidences are cited and consult follow assessor and discuss if the assigned scores could be compared and discuss the assigned scores and the rational pertained to the scores by providing evidences and arguments and determine whether to hold on to the original score or make adjustments.

The measurement of teacher competences on five Likert scales of 65 item TBSC led to the scores ranging from a minimum score of 4 to a maximum score of 34, and the results were put into SPSS to investigate the reliability and validity of the inventor. In the inter-rater reliability of raters, the rating was reported in the previous section.

Results

The purpose of the present study is to design and validate a teacher inventory called the teacher balanced scorecard (TBSC) by computing its reliability and validity—both exploratory and confirmatory methods in order to enable researchers to employ it in their future studies. The TBSC questionnaire includes 65 items which measure student, departmental, and learning perspective each of which has four aspects. The data were analyzed in order to probe its reliability and exploratory and confirmatory factor analyses. Before discussing the results, it should be mentioned that the assumptions of univariate and multivariate normality were met. As noted by Bae and Bachman (2010), the absolute values of the skewness and kurtosis values (Table 3) were lower than 1.96, indicating univariate normality of the data.

Table 3 Testing univariate and multivariate normality assumptions

The multivariate normality assumption was also retained. The Mardia index of .009 was lower than ± 3 (Bae and Bachman 2010).

Cronbach’s alpha reliability indices

Table 4 displays the Cronbach’s alpha reliability indices for the three main perspectives and their dimensions. The reliability indices for the student, departmental, and learning perspectives were .90, .91, and .76, respectively. The latter had only 10 items. The reliability indices of the dimensions ranged from a low of .65 for personal aspect of learning which had only two items to a high of .93 for the technical aspect of departmental perspective.

Table 4 Reliability statistics

Exploratory factor analysis

A factor analysis was run to probe the underlying constructs of the 65 items of the TBSC questionnaire. Figure 1 suggested 3 to 12 factors to be extracted. The 12 extracted factors accounted for 53.25% of the total variance. Since the TBSC questionnaire had 12 subsections, it was decided to extract the 12 factors using principal axis factor method and varimax rotation. The 12 extracted factors accounted for 53.25% of the total variance.

Fig. 1
figure 1

Optimum number of factors proposed by SPSS

Table 5 displays the factor loadings of the 65 items under the extracted factors. Based on these results, it can be concluded that:

  • The first factor includes the 16 items related to the technical aspect of departmental perspective.

  • The 10 items related to the personal aspect of student perspective loaded under the second factor.

  • The 10 items loading under the third factor were related to the clinical aspect of student perspective.

  • The fourth factor includes the 7 items related to the technical aspect of student perspective. Item 28, which was the single indicator of the critical aspect of student perspective, also loaded under the fourth factor.

  • The 4 items related to the critical aspect of learning perspective loaded under the fifth factor.

  • The 4 items loading under the sixth factor were related to the critical aspect of departmental perspective.

  • The seventh factor includes the 4 items related to the clinical aspect of departmental perspective.

  • The 3 items related to the personal aspect of departmental perspective loaded under the eighth factor.

  • The 2 items loading under the ninth factor were related to the technical aspect of learning perspective.

  • The tenth factor includes the 2 items related to the clinical aspect of learning perspective, and finally,

  • The 2 items related to the personal aspect of learning perspective were loaded under the second factor. The 12th factor did not include any meaningful (≥ .30) loadings.

Table 5 Rotated factor matrix

Based on these results, it can be concluded that the construct validity of the TBSC questionnaire was confirmed employing an exploratory method.

Confirmatory factor analysis

Four separate structural equation models (SEMs) were developed and tested in order to probe the trait structure of the TBSC questionnaire. The three SEM models targeted the three perspectives individually, while the last model explored the structure of the total data.

Confirmatory factor analysis of student perspective

The trait structure of the three components of the student perspective is displayed in Fig. 2. Except for the critical aspect which was dropped from the model, the figure shows the standardized relationships between the items (blue squares) and their related aspects (yellow ovals) which eventually contributed to the “student” perspective (green oval).

Fig. 2
figure 2

Trait structure of student perspective

All items have significant contributions to their respective dimensions (≥ .30), and all three aspects also significantly loaded on the student perspective. The non-significant chi-square statistics (χ2 (321) = 351.67, p = .115) indicated that the model enjoyed a good fit. The ratios of the chi-square over the degree of freedom, i.e., 351.67/321 = 1.09, was lower than 3. These results also supported the fit of the model. The RMSEA statistic and its 90% confidence intervals (RMSEA = .021, 90% CI [.000, .034]) were lower than .05.

The model enjoyed a good fit. The PCLOSE statistic of one was higher than .05. All these statistics proved the fit of the model. The indices of NFI, NNFI, CFI, IFI, and RFI were all higher than .90, indicating fit of the model. The critical N (CN) value of 235.07 was higher than 200. The CN results proved the sampling adequacy of the present model. Table 6 displays the fit indices related to the student perspective.

Table 6 Fit indices; student perspective

Confirmatory factor analysis of departmental perspective

The trait structure of the three components of the departmental perspective is displayed in Fig. 3. Although the chi-square statistic was significant (χ2 (320) = 370.79, p = .026), it indicated that the model did not enjoy a good fit. Since the chi-square statistic is sensitive to large sample sizes, its ratio over the degree of freedom should be consulted. The PCLOSE statistic of .24 was higher than .05. All these statistics proved the fit of the model.

Fig. 3
figure 3

Trait structure of departmental perspective

All items have significant contributions to their respective dimensions (≥ .30), and all four aspects also significantly loaded on the departmental perspective. The ratios of the chi-square over the degree of freedom; i.e., 370.79/320 = 1.15, was lower than 3. These results also supported the fit of the model. The RMSEA statistic and its 90% confidence intervals (RMSEA = .027, 90% CI [.010, .039]) were lower than .05.

The indices of NFI, NNFI, CFI, IFI, and RFI were all higher than .90, indicating fit of the model. The critical N (CN) value of 222.37 was higher than 200. The CN results proved the sampling adequacy of the present model. Table 7 displays the fit indices related to the departmental perspective.

Table 7 Fit indices; departmental perspective

Confirmatory factor analysis of learning perspective

The trait structure of the three components of the learning perspective is displayed in Fig. 4.

Fig. 4
figure 4

Trait structure of learning perspective

The PCLOSE statistic of .24 was higher than .05. All these statistics proved the fit of the model.

All items have significant contributions to their respective dimensions (≥ .30), and all four aspects also significantly loaded on the learning perspective. The model enjoyed a good fit, although the chi-square statistic was significant (χ2 (31) = 55.20, p = .004) indicated that the model did not enjoy a good fit. Since the chi-square statistic is sensitive to large sample sizes, its ratio over the degree of freedom should be consulted. The ratios of the chi-square over the degree of freedom, i.e., 55.20/31 = 1.78, was lower than 3. These results also supported the fit of the model. The RMSEA statistic and its 90% confidence intervals (RMSEA = .060, 90% CI [.033, .086]) were between .05 and .08. This range is considered as “reasonable fit” by Byrne (2016).

The indices of NFI, NNFI, CFI, IFI, and RFI were all higher than .90, indicating fit of the model. The critical N (CN) value of 204.31 was higher than 200. The CN results proved the sampling adequacy of the present model. Table 8 displays the fit indices related to the learning perspective.

Table 8 Fit indices; learning perspective

Confirmatory factor analysis of learning perspective

The trait structure of the three components of the TBSC overall model is displayed in Fig. 5.

Fig. 5
figure 5

Trait structure of TBSC overall model

All aspects had significant contributions to their respective dimensions (≥ .30), and all three perspectives also significantly loaded on the TBSC. The model enjoyed a good fit, although the chi-square statistic was non-significant (χ2 (41) = 26.72, p = .958) indicated that the model enjoyed a good fit. The ratios of the chi-square over the degree of freedom, i.e., 26.72/41 = .65, was lower than 3. These results also supported the fit of the model. The RMSEA statistic and its 90% confidence intervals (RMSEA = .000, 90% CI [.000, .000]) were all lower than .05 and indicated that the present model enjoyed a good fit.

The PCLOSE statistic of one was higher than .05. All these statistics proved the fit of the model. The indices of NFI, NNFI, CFI, IFI, and RFI were all higher than .90, indicating fit of the model. The critical N (CN) value of 523.62 was higher than 200. The CN results proved the sampling adequacy of the present model. Table 9 displays the fit indices related to the overall model.

Table 9 Fit indices; TBSC overall model

Discussion

This study was an attempt to design a teacher evaluation inventory named TBSC which focuses on teacher competences from the three perspectives of student, departmental, and learning and growth ones. Cronbach’s alpha reliability indices for the three main perspectives and their dimensions show that the assessment is instrument independent (good internal consistency). The results of the exploratory factor analysis indicated that there was no construct-irrelevant factor, and all the indicators were loaded in the related teacher competence and perspective dimension and they assess what they are supposed to asses. Four separate structural equation models (SEMs) were tested in order to probe the trait structure of the TBSC questionnaire. The first three SEM models targeted the three perspectives individually, while the last model explored the structure of the total data. The results indicated that all items had significant contributions to their respective dimensions. This study can provide insights into how to manage the criticism made to teacher value-added approaches towards teacher education and evaluation.

Werbińska’s (2015) review of most approaches towards teacher education highlighted the fact that most appraisal systems are product based which do not provide any information of the teaching reforms taking place in teaching development. He criticizes the teacher induction programs since they are output-based and focuses on teacher certification or student achievements. Later observation-based checklist tick off points which leave the understanding of the context of the observation behind decreases the value of teacher education programs. Mentee observation feedback on critical incidences also leaves no space for teacher themselves to evaluate their own act of teaching. The use of artifacts such as running commentaries and transcribed feedback sessions was introduced as catalysis. The fact is that all the approaches have something in common and that is teachers reform their teaching on the basis of the mentor or supervisors which prevent teachers from forming their own identity. Therefore, this sets a divide between advocators of student-centered progressivist and teacher authority-based defensive teaching. This inventory can direct teachers in their self-evaluation and reflection and help teacher transformation. A transformation which entails a change from a mastery teaching is giving priority to the appropriate act of delivering teaching to learners to what Richards (2010) calls “learner-focused teaching.” It is a kind of teaching in which the focus is maximizing the potential for learning.

To be more specific with this inventory and its efficacy in teacher evaluation, each competence indicators are reviewed. The first competence is critical competence which requires teachers being engaged with teamwork, maximizing teaching quality via asking and suggesting critical ideas, volunteering in policy and program making tasks and initiating actions, and sharing innovations and developments. These indicators are in line with what Richards (2010) requires all teachers to develop—pedagogical content knowledge. He distinguishes pedagogical content knowledge from disciplinary knowledge. Disciplinary knowledge is the teachers’ knowledge of his discipline. In the case of linguistics, it can be the knowledge of semantics, syntax, and discourse and pragmatics. Whereas pedagogical content knowledge is the knowledge about teaching and learning which helps the teachers to solve problems raised in the actual classroom context, and it is the knowledge acquired through reflective thinking. Reflection includes looking back and forward to teaching experiences and initiating necessary changes and managing the consequences of those changes. As Mezirow (2000) suggests, reflection should be both on content (teaching experiences) and process (how the problems are solved and ongoing development are achieved).

The other competence is personal competence. The indicators of personal competence include personal involvement and establishing a sense of community. Part of teacher development comes from participating with communities having the same goal, interests, and values. The sense of community creates collegiality which provides opportunities for group-oriented activities and joint problem solving, and this helps the learners to play new roles of team leaders, teacher trainer, and mentor and critical friend (Richards and Farrell 2005).

The other competence is clinical competence. The indicators of these competences show that they are related to real-time teaching action. The indicators suggest learner-focused teaching. The development of these competences shows how teachers transit from a survival and mastery stage to a stage where teachers focus more on learners’ learning. At survival stage, teachers act within their comfort zone and focus on their teaching, and at later stages of development, they focus more on the impact of their teaching on student learning (Farrell 2012). The trend of change in critical competence to technical and clinical competence indicates that when mind undergoes changes as a result of reflection through a portfolio, its results can be seen in actions in classes. A transformation which entails a change from a mastery teaching which is giving priority to the appropriate act of delivering teaching to learners to what Richards (2010) calls “learner-focused teaching.” It is a kind of teaching in which the focus is maximizing the potential for learning. Besides learner-focused teaching, teachers gain skills in reasoning, application of pedagogical content knowledge (a knowledge by which they can manage their teaching), anticipate and recognize problems, and take actions for solving them. Besides, as teachers increase their knowledge and experience, they develop improvisational teaching which is moving towards flexibility in teaching. Improvisational teaching is having cognition behind the teaching skills acquired through experiences (Richards 2010).

The other competence is technical competence. A review of indicators of technical competence shows that this competence is related to the metacognition and pre- and post-planning of teaching act. Teachers need to be aware of the teaching they do which means teachers should develop professionalism. Professionalism is to be technical both at large-scale dimension responding to institutionally prescribed teaching to be accountable in terms of managerial dimensions pertained to ministries of education and teaching organizations and local scale dimension which is called independent professionalism and requires teachers to be consciously aware of ones’ teaching practices.

Conclusion

This study aimed at designing and validating an instrument for teacher evaluation. The indicators pertained to four teacher competences including critical, clinical, technical, and personal competences were identified theoretically through literature review and operationally through focused group discussion by the panel of experts. The results of reliability analysis indicated that the assessment on the basis of the newly developed inventory is instrument independent which means that the inventory enjoys internal consistency. The results of the exploratory factor analysis indicated that there was no construct-irrelevant factor, and all the indicators were loaded in the related teacher competence and perspective dimension and they assessed what they were supposed to asses.

The results of this study are of great significance for education research and teacher development. This inventory makes the teachers capable of monitoring their class automatically. The TBSC inventory can also help teachers self-evaluate their teaching ability and performance. TBSC inventory can help teachers to monitor their teaching timely and dynamically. The items of TBSC (can be checked in Table 1 and Additional file 1) are related to teachers’ awareness in optimizing teaching quality. Besides, TBSC inventory can help teachers identify their strength and weakness and track their learning and growth.

However, certain caveats apply to the conclusions. First, the sample consists of EFL teachers in one of the districts of Iran, and hence, we cannot claim that the inventory has the same potential in all educational contexts which limits the generalizability of its use. Second, it might be possible that different educational context shows a different compliance with the inventory because education ideology is fostered in that context. Educational systems are primed with certain ideologies that mediate any changes happening in the education ecology. These macro-ideologies are the co-creative and directors of teacher perception and practices (Vasileiadis et al. 2013) that act as filters that legitimize serotypes in teacher perception and teacher practices (Vasileiadis et al. 2013). The macro ideologies have implicit messages for teachers. For example the anti-American attitude Iranian government instigates may imply that EFL teachers’ development is not appreciated since learning that the target language may bring the values of the target culture to native one or the government’s lack of infrastructure facilities to implement technology-mediated learning and teaching may imply that whether teachers keep up with the latest education technology is necessary and cause teachers not take technology serious in their teaching practice. The invention or implementation of any inventories without paying attention to its origin and context that is formed may jeopardize what Cohen (1995) calls coherence in practice. For any attempts, there should be coherence between teachers and education ecology. Future research can implement the resulting teacher competence inventory in different teaching context to attest the accreditation of this inventory in different educational contexts since teaching is liable to the charge of different biases and understandings that are culture-bound.