Abstract
Competency, which has complexity, means that competency assessment has to be done elaborately and indepth. The importance of assessment is that it has to show the evidence of validity. Is assessment of performance competency by the methods of rash analysis, inter-subtest correlation analysis, and confirmatory factor analysis valid or not. The purpose of research is to examine the validity of assessment of performance competency by the methods of rash analysis, inter-subtest correlation analysis, and confirmatory factor analysis. The sample group consisted of directors and employees. The instruments included three volumes - assessment by director (Volume I), self-assessment (volume II), and supervisor assessment (volume III). The data were analyzed by the rash analysis, inter-subtest correlation analysis, and confirmatory factor analysis. Results revealed that the examination by the rash analysis volume I fit almost all of the items, whereas volume II and volume III had all values in the range of statistical acceptance, the inter-subtest correlation analysis in all volumes had a high level of relationship, and the confirmatory factor analysis for all volumes was concordant with the empirical data. The evidence of examination on three methods based on the measurement standard for item response theory and classical test theory can indicate the quality of assessment of performance competence very well.
Keywords: ValidityPerformance competency;Rash analysisInter-subtest correlation analysisConfirmatory factor analysis
1. Introduction
Society is always changing; therefore, a successful organization has to be improved and developed
all the time. Organizational success depends on the drive of human resources. The organization which
has the ready, potentiality, and competent human resources for performance gets the advantage and has
the opportunity to achieve great success. Hence, the human factor is considered crucial factor for the
performance in the organization. The success of an organization, thus, depends on employee
performance as the crucial factor. The indicator leading to the success is essential for performance of
employees in the organization (Kesler, 1995; Beheshtifar, 2011). An individual will expose any
competency and usually depends on various components in terms of knowledge, skills/competency, and
other characteristics. These crucial components affect the competency of individuals (behavioral
characteristics of performance). Moreover, the competency also has a relationship with the
accomplishment (Office of the Civil Service Commission, 2005).
Competency is complicated; hence, the competency assessment should be done elaborately, deeply,
and with variation. This is in accordance with the concept of Baartman et al. (2006) who said that the
competency assessment is very complex, and that one single assessment method seems not to be
sufficient. Moreover, Koeppen et al. (2008) suggested that the competencies are conceptualized as
complex ability constructs. For the challenge point of this research about competency assessment, one
of the points is about psychological models (Frey & Hartig, 2009). In addition, Vazirani (2010)
represented the challenge of using competency models for measuring or appraising certain areas of
performance and providing developmental feedback based on these assessments. Thus, it can be seen
that competency is complex. It has to be assessed elaborately and with in-depth scrutiny. Only a single
method cannot be sufficient, and the interested point to be studied should be in terms of psychological
models.
Competency assessment has to be done indirectly; it cannot be measured directly but has to use
qualified instruments, and the qualified instruments have to be considered carefully in terms of validity.
According to the Educational Testing Service, ‘validity is the most important aspect of the quality of an
assessment’ (2002), and the American Educational Research Association et al. stated that ‘validity is
the most fundamental consideration in developing and evaluating tests’ (1999) According to the
definition of validity, from the Standards for Educational and Psychological Testing (as cited in AERA
et al., 1999), validity refers to “the degree to which evidence and theory support the interpretations of
test scores entailed by proposed uses of tests”. The validation process, therefore, involves the
accumulation of evidence to support the proposed test score interpretations and uses. The process of
accumulating evidence to support the validity of test score interpretations starts prior to the
development of an assessment (AERA et al., 1999). This suggests that the evidence of validity on the
instruments used to assess is the qualification which had to be considered at the first stage.
Regarding the backgrounds of this research, in terms of competency which had complexity,
competency assessment has to be done elaborately and with in-depth scrutiny. The importance of
assessment is that it has to show the evidence of validity by correct and accurate evaluation; thus,
various methods of assessment are necessary to be used in order to reach correctness and accuracy of
competency assessment. This will be the reflection of quality through competency assessment to lead to
measurement, assessment, and development of personnel for performance in the organization
effectively and successfully with sustainable performance.
2.Purposes of the study
The purpose of this research is to examine the validity of assessment of performance competency by
the methods of rash analysis, inter-subtest correlation analysis, and confirmatory factor analysis.
3.Research methods
3.1. Participants
1) The group of employees in the self-assessment is the TISCO company’s employees working
between 2013 and 2014. Tisco is a financial institute running the business with the highest standard of
good supervision. The Company is considered as a large company, consisting of Tisco Financial Group,
Tisco Bank, Tisco Securities, and Tisco Asset. The Company has been numerously rewarded for
success, which reflects the quality of the organization, such as the reward of Board of The Year 2013,
Top Corporate Governance Report Awards 2013, Outstanding Securities Company Awards 2013 –
Retail Investors, Outstanding Securities Company Awards 2013 – Institutional Investors, Outstanding
Investor Relations Awards 2013, “A” on Hewitt Best Employers in Thailand 2013, and Top Bank in the
Secondary Market for Corporate Bonds 2012 (Tisco, 2014). The subjects are from stratified random
sampling by having the strata pillar, as well as having the employees as the random unit from the
assessed employees for 3,929 persons.
2) The group of assessments by directors were obtained from purposive sampling. They are the
direct superiors of the employees.
3) The group of employees in the supervisor assessment consists of 516 TISCO Company’s
supervisors from stratified random sampling by having the strata pillar.
3.2. Instrumentation
content validity, by consideration of the experts, has the value of index of consistency between .50 and
1.00. The form is used by the employees who have to assess 52 supervisors. The value of item-total
correlation was between .22 and .53. The reliability was carried out by using the formula of Cronbach's
alpha coefficient equal to .93 and equal to Spearman-Brown coefficient .83.
3.3. Procedure
For the process of validity examination on instruments of three volumes, each volume is examined
by three methods: 1)Rash analysis — this method will examine the fundamental data of the uni-
dimensionality or multi-dimensionality by factors analysis considered by Eigen value, then it will be
analyzed by multi-dimensional random coefficients multinomial logit model (MRCMLM), which is
considered by INFIT mean square or weighted mean square (INFIT MNSQ), 2) Inter-subtest
correlation analysis — this can be analyzed by the relationship of five sub-competency scores and total
competency scores by using the Pearson correlation coefficient considered by the validity from
statistical significance, and 3) Confirmatory factor analysis — this is the confirmation of five
competency components considered by the concordance of the model with empirical data from the chi
square, goodness of fit index, adjusted goodness of fit index, root mean square residual, and root mean
square error of approximation
4.Findings
4.1. Rash analysis
4.1.1. Fundamental data examination
The analysis results of performance competency dimension by factors analysis was found that the
ratio between the Eigen value of factor 1 toward the Eigen value of factor 2 of the assessment by
directors was equal to 1.315, the self-assessment was equal to .001, and the supervisor assessment was
equal to 1.652. The values were less than the criteria, which would be concluded by the
unidimensionality. Morizot et al. (2007) proposed to consider the unidimensionality from the ratio
between Eigen value of factor1 toward the Eigen value of factor 2. If the value was more than or equal
to 3.00, it indicated the unidimensionality. With all evidence mentioned above, it can be concluded that
the performance competency assessed by the assessment by directors, self-assessment and supervisor
assessment were appropriate for multidimensionality assessment.
4.1.2. Research result of multidimensionality
According to this research, the researchers studied the base model called multidimensional random
coefficients multinomial logit model (MRCMLM) (Adams et al.,1997). Because there are five
competency assessment methods, with their fundamental characteristics about multi-dimensionality, it
fits with the analysis by the multidimensional form of the partial credit model. This is the model that
fits to the measurement having multi values response, personality measurement, metacognition and
attitudes measurement, and it fit to the measurement using the items of giving multi values scores by
ordered polytomous items (Ostini, & Nering, 2006). The data was analyzed by the ConQuest 2.0
computer program. The analysis results were concordant between the assessment model of performance
competency and question items considered by the INFIT mean square or weighted mean square (INFIT
MNSQ). These are the values based on the root mean square error (RMSE) concept between the
estimated and true item response (DeMars, 2004). The statistical values indicated that the data of each
item is concordant with the item response model determined, and does not accept the value between .60
and 1.40, which is the criteria for consideration of value measurement by rating scale (Wright et al.,
1994). The analysis result shown in Table
MNSQ value between .79 and 1.45. Only a single question item was beyond the acceptance range,
which was, COM2. Whereas the self-Assessment had the INFIT MNSQ value between .98 and 1.01,
and the supervisor assessment had the INFIT MNSQ value between .88 and 1.09, which its value was
in the acceptance range. This evidence shows the validity according to the theoretical construct of the
assessment of performance competency.
4.2. Inter-subtest correlation analysis
The construct validity is investigated by the inter subtest correlation method between the
dimensional or sub-test scores and total-test scores (Kline et al., 1994; Cohen, & Swerdlik, 2005; 2010;
Xi, 2008; Canivez et al., 2009). By using the Pearson correlation coefficient with the volume I,
assessment by director had the value between .675 and .940. The competency which had the highest
relationship with the total scores was LOG, secondly was BUS, LEA, COM, and PEO, respectively.
For the volume II, the self-Assessment had the value between .441 and .650. The competency which
had the highest relationship with the total score was LOG, secondly was COM, BUS, LEA, and PEO,
respectively. And the volume III, the supervisor assessment, the value was between .519 and .884. The
competency which had the highest relationship with the total score was LOG, secondly was BUS, LEA,
COM, and PEO, respectively. All values of all volumes had the statistical significance at .01 level, and
the relationship was at a high level. This evidence shows the validity according to the theoretical
construct of the assessment of performance competency as shown in Table
4.3. Confirmatory factor analysis
According to the analysis result of the models on assessment of performance competency, five major
indicators were found. These consisted of communication, people, logic, business and leadership within
the three volumes of the assessment form, When examining the concordance of performance
assessment models and empirical data, it was found that the assessment by director volume had the chi
square equal to 1.93 (df=2, p=.38), because the chi square was sensitive to the size of sample group.
Therefore, it had to mutually consider other statistical values. The goodness of fit index (GFI) was
equal to 1.00, the adjusted goodness of fit index (AGFI) was equal to 1.00, the root mean square
residual (RMR) was equal to .00, and the root mean square error of approximation (RMSEA) was equal
to .00. The self-assessment had the chi square = 5.65 (df = 2, p = .059), GFI = 1.00, AGFI = 1.00, RMR
= .00, RMSEA=.02, and the supervisor assessment had the chi square = 4.30 (df = 4, p = .37), GFI =
1.00, AGFI = .99, RMR = .01, RMSEA=.01. When considering the data according to the assessment
criteria of Bollen (1989), Kelloway (1998), and Hair et al. (2010) from the fit index value of the model,
the data showed that the assessment model of performance competency was concordant with the
empirical data. This evidence shows the validity according to the theoretical construct of assessment of
performance competency.
5. Conclusion
The validity examination result of assessment of performance competency was carried out by using
five competencies, which consisted of communication, personnel, logic, business, and leadership. The
examination was done by using the rash analysis method, which found that the volume I, assessment
by director, had the INFIT MNSQ value between .79 and 1.45 from all twenty-four items. Only a single
question item was beyond the acceptance range. For the volume II, self-assessment method, this had the
value of INFIT MNSQ between .98 and 1.01, and the volume III, supervisor assessment had the value
of INFIT MNSQ method between .88 and 1.09 which had the value in the range of statistical
acceptance. The examination by inter-subtest correlation analysis was found that the volume I had the
value between .675 and .940, the volume II had the value between .441 and .650, and the volume III
had the value between .519 and .884. The competency which had the highest relationship with the total
scores of all volumes was LOG. All values, of all volumes, had the statistical significance at .01 level,
and the relationship value was in a high level. For the examination that used the confirmatory factor
analysis method, it was found that the assessment model of performance competency of all volumes
was concordant with the empirical data. According to the evidence of examination on the three methods
based on the measurement standard for item response theory and classical test theory, these can indicate
the quality of assessment of performance competence very well.
6. Discussion
Regarding the instruments used for assessment created by the five aspects of competency, which are,
communication, personnel, logic, business, and leadership, for three volumes, consisting of the
assessment by director (Volume I), self-assessment (volume II), and the supervisor assessment (volume
III), the validity of these instruments were examined by three methods, which included the rash
analysis, inter-subtest correlation analysis, and confirmatory factor analysis.
The rash analysis results by considering the item fit value based on the concept of root mean square
error (RMSE) between estimated and true item response (DeMars, 2004) is the statistics value which
indicates whether the item data are concordant with the item response model. The item data, concordant
with the model, suggested that item has the response form concordant with the model in a form of
cumulative frequency curve (Chianchana, 2009), while only one question item was beyond the
statistical acceptance of Wright and others, 1994, which is, assessment by director (Volume I). Whereas
the question item of Com2 is “plan of representation by expecting that it can motivate others”, this
question item is still not concordant with the construct according to the item response theory; however,
this is considered as a small abnormality from the response structure, while other volumes showed that
every question item was concordant with the theoretical construct.
The result of inter-subtest correlation analysis for all volumes showed that all values had a highly
positive relationship. This was the evidence shown between the dimensional or sub-test scores and
total-test scores. These highly internal relationships reflect the validity of all series of assessment forms
developed under the same theory, considered as the aspect of homogeneity (Kline et al., 1994; Cohen,
& Swerdlik, 2005; 2010). All volumes have the logic competence related highly with the total
competence score. For the measured logic, it is the concept involved with the reason giving, and
reasonable reason criteria. The indicator consists of basic analysis, basic analysis and synthesis,
complex situations, and holistic and systematic. For the logic characteristic, Copi (2002) said that logic
is about reasoning on every subject — science and medicine, ethics and law, politics and commerce,
sports and games, and even the simple affairs of everyday life — which are considered as crucial
factors towards the performance.
The result of confirmatory factor analysis for all volumes showed that the assessment model was
concordant with the empirical data. The confirmatory factor analysis was then employed so that the
researcher could test the relationship between the model and the empirical data (McIntire and Miller,
2007). This means that all five structures of competency assessment forms were valid by the
assessment of structural models through the statistical index including chi-square test statistics,
goodness-of-fit index (GFI), adjusted goodness-of-fit index (AGFI) , root mean squared residual
(RMR) , and root mean squared error of approximation (RMSEA). All indexes complied with a
consideration of Bollen (1989), Kelloway (1998) and Diamantopoulos and Siguaw (2000), who showed
the model congruence with empirical data (Bollen, 1989; Kelloway, 1998; Diamantopoulos, & Siguaw,
2000). Moreover, when the assessment of measurement model, through the consideration from the
value of indicator loading, which is the validity evidence (Diamantopoulos, & Siguaw, 2000), it
appeared that all loading volumes have statistical significance, which supports the validity of all
instruments for competency assessment. The crucial competency, when considering the loading value
in the assessment by director (Volume I) and supervisor assessment (volume III), is about the logic.
This is in accordance with the examination result of validity, by inter-subtest correlation analysis,
whereas the self-assessment (volume II) contradicts the result of inter-subtest correlation analysis.
Another crucial competency is communication, which is the personal characteristic which has fluency,
persuasion, presentation, and negotiation. These characteristics are crucial, but not more or less than the
logic. This is in accordance with the requirement of employers who need the employees who have
communication competency and presentation skills (Woodward et al., 2010). In addition, this is in
accordance with Kelley and Bridges’ study (2005), which showed that business communication skills
and presentation skills are the top two of twenty-three skills that are necessary for a successful career.
7. Recommendation
The instruments for competency assessment of performance have various volumes, and they are
conducted under the organizations which can control the examination effectively. For further research,
it should apply the technology to help originate the quickness, and saving. Moreover, it should be
created to be appropriate to the responders, until it builds the sustainability of the test. The conduction
should be done in terms of computer adaptive testing (CAT). Moreover, the nature within organizations
to have various groups of employees means that the instruments used for assessment should be
evaluated carefully to consider these differences of variance. For further research, the differential item
functioning (DIF), and differential test functioning (DTF) should be examined to show the evidence of
quality on the tests and questionnaires in terms of fairness through every group of responders.
References
- Adams, R.Wilson, M.Wang, W. (1997). The multidimensional random coefficients multinomial logit model. Applied Psychological Measurement, 21, 1-23
- American Educational Research Association., American Psychological Association., & National Council on Measurement in Education., (1999). Standards for educational and psychological ed.). , testing. (2nd
- (), Washington, DC: American Educational Research Association.
- Baartman, L. K. J.Bastiaens, T. J.Kirschner, P. A.Van Der Vleuten, C. P. M. (2006). The wheel of competency assessment: Presenting quality criteria for competency assessment programs. Studies in Educational, 153-170, Evaluation. 32
- Beheshtifar, M. (2011). Role of career competencies in organizations. European Journal of Economics, Finance and Administrative Sciences, 42, 6-12
- Bollen, K. A. (1989). Structural equations with latent. , variables. New
- Chianchana, C. (2009). Multidimensional analysis. Journal of Education Khon Kaen University, 32(4), 13-22
- Cohen, R. J.Swerdlik, M. E. (2005). Psychological testing and assessment: An introduction to tests and measurement (6th ed.). Boston: McGraw-Hill.
- Cohen, R. J.Swerdlik, M. E. (2010). Psychological testing and assessment: An introduction to tests and measurements (7th ed.). Boston: McGraw-Hill.
- Copi, I. M. (2002). Introduction to logic (11th ed.). New Jersey: Pearson Education.
- DeMars, C. (2004). Measuring higher education outcomes with a multidimensional rasch model. Journal of Applied Measurement, 5(4), 350-361
- Diamantopoulos, A.Siguaw, J. A. (2000). Introducing LISREL: A guide for the SAGE Publication. , uninitiated. London
- Educational Testing Service (ETS)., (2002). ETS standards for quality and fairness.
- Frey, A.Hartig, J. (2009). Assessment of competencies. Studies in Educational Evaluation, 35, 55-56
- Hair, J. F.Anderson, R. E.Tatham, R. L.Black, W. C. (2010). Multivariate Data Analysis: A Global Perspective (7th ed.). New Jersey: Pearson education.
- Kellowey, E. (1998). Using LISREL for structural equation modeling: A researcher's guide.
- Kelly, C. A.Bridges, C. (2005). Introducing professional and career development skills in the marketing curriculum. Journal of Marketing Education, 27(3), 212-218
- Kesler, G. C. (1995). A model and process for redesigning the HRM role, competencies, and work in a major multi-national Resource Management. , 34, 229-252, corporation. Human
- Kline, R. B.Snyder, J.Gulmette, S.Castellanos, M. (1994). Evaluation of the construct validity of the kamphaus-reynolds supplementary scoring system for the K-ABC. Assessment, 1(3), 219-226
- Koeppen, K.Hartig, J.Klieme, E.and Leutner, E. (2008). Current issues in research on competency modeling and assessment. Zeitschrift fur Psychologie/Journal of Psychology, 216, 60-72
- McIntire, S. A.Miller, L. A. (2007). Foundations of psychological testing: A practical approach (2nd ed.).
- Morizot, J.Ainsworth, A. T.Reise, S. (2007). Toward modern psychometrics: Application of item response theory models’, In, R.C. Fraley, & R.F. Krueger. (Eds.), Handbook of Research Methods in Personality Psychology. , 407-423, R.W. Robins
- Office of the Civil Service Commission., (2005). Competency manual of civil service. , P.A. living
- html [29 September 2014] Vazirani, N. (2010). Competencies and competency model-A brief overview of its development and Journal of Management. , 7(1), 121-131, application. SIES
- Woodward, B.Sendall, P. (2010). Integrating soft skill competencies through project-based learning across the information systems curriculum. Information Systems Education Journal, 8(8), Retrieved from http://isedj.org
- Wright, B. D.Linacre, J. M.Gusafson, J. E.Martin-Lof, P. (1994). Reasonable mean-square fit values.
- htmXi, X. (2008). Methods of test E. Shohamy, and N.H. Hornberger. (Eds.), Encyclopaedia of Languate and Education: Vol.Language Testing and Assessment (pp.. 177-196), (2nd ed.). , 7, validation. In
- (), 212
Copyright information
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
About this article
Publication Date
14 May 2016
Article Doi
eBook ISBN
978-1-80296-007-5
Publisher
Future Academy
Volume
8
Print ISBN (optional)
-
Edition Number
1st Edition
Pages
1-252
Subjects
Psychology, social psychology, group psychology, collective psychology, teaching, teaching skills, teaching techniques
Cite this article as:
Chianchana, C., & Na Wichian, S. (2016). Validation of the Assessment of Performance Competency. In Z. Bekirogullari, M. Y. Minas, & R. X. Thambusamy (Eds.), Cognitive - Social, and Behavioural Sciences - icCSBs 2016, May, vol 8. European Proceedings of Social and Behavioural Sciences (pp. 203-212). Future Academy. https://doi.org/10.15405/epsbs.2016.05.21