Background/rationale
Assessments and psychometric models developed under the trait psychology perspective have been reliable methods for evaluating the general state of students’ knowledge, skills, and abilities when the purpose of measurement is to compare students’ abilities and to select students who have developed mastery in the context of a licensing examination. However, overall scores of this type do not offer sufficient useful information for the purposes of (1) measuring multidimensional contextual knowledge, skills, and abilities; (2) measuring complicated tasks reflecting complex knowledge, skills, and abilities; (3) understanding distinguishable systematic patterns associated with different characteristics of groups; and (4) providing diagnostic information connected with the curriculum and instruction. For these purposes, it is necessary to obtain more information from assessment results through various measurement models.
More specifically, the main purpose of a large-scale assessment is to compare students’ achievement levels and to make pass/fail decisions based on general proficiency. This is usually done by students’ overall scores, which are used to assign students to specific performance levels. However, this information has very little instructional usefulness in terms of what should be done to improve individual students’ levels of achievement. That is, overall test scores from large-scale assessments offer relatively little diagnostic information about a student’s strengths and weaknesses [
1]. Diagnostic information is more informative for students, instructors, and assessment/instruction developers from the perspective of learning and improving the quality of assessments and the curriculum [
2]. In light of these issues, diagnostic classification models (DCMs) have been proposed as psychometric models.
DCMs are statistical models that were originally developed to classify students in terms of their mastery status for each attribute [
3,
4]. DCMs contain multiple attributes, which refer to latent aspects of knowledge, skills, and abilities that are supposed to be measured in an assessment. Students’ mastery status for the attributes of interest are estimated based on their observed response patterns. A composite of a student’s mastery statuses for the attributes is referred to as an attribute profile. Therefore, the attribute profile is a pattern used for providing diagnostic feedback. Several DCMs have been proposed, such as deterministic inputs, noisy “and” gate (DINA), deterministic inputs, noisy “or” gate (DINO), and the re-parameterized unified model.
These models differ depending on the variables of interest and the condensation rules that are used for modeling attributes; however, a central concept of modeling is linking the diagnostic classification with cognitive psychological findings [
4]. Since multiple attributes are involved and tasks can depend on more than one attribute, their relationships are represented by a complex loading structure, often called a Q matrix [
4]. A Q matrix contains the targeted attributes and specification of which attributes are measured by which task(s) based on substantive theoretical input (e.g., a domain specialist for the relevant examination). To construct a Q matrix, many sources may be used, such as subject matter experts’ opinions, cognitive developmental theory, learning science, and learning objectives in the curriculum.
Educators often want diagnostic information from assessments, and in particular, educators in the health professions often want to provide feedback to a given student based on how he or she does on each content strand. However, most assessments are developed to provide only a single total score [
3]. Most score reports in the health professions provide a total score or pass-fail decisions based on classical test theory. DCMs are psychometric models that characterize examinees’ responses to test items through the use of categorical latent variables that represent their knowledge [
5]. Thus, DCMs have become a popular area of psychometric research. However, few application studies using DCMs with health professions data have been reported [
6]. The purpose of this study was to conduct a DINA analysis using Korea Health Professional Licensing Examination data in order to provide diagnostic information about each content domain in this licensing examination.