Skip Navigation
Skip to contents

JEEHP : Journal of Educational Evaluation for Health Professions

OPEN ACCESS
SEARCH
Search

Articles

Page Path
HOME > J Educ Eval Health Prof > Volume 4; 2007 > Article
Review
Reconsidering the Cut Score of Korean National Medical Licensing Examination
Duck Sun Ahn*, Sowon Ahn

DOI: https://doi.org/10.3352/jeehp.2007.4.1
Published online: April 28, 2007

Department of Medical Education, College of Medicine, Korea University, Seoul, Korea

*Corresponding e-mail: dsahn@korea.ac.kr

: 

• Received: February 10, 2007   • Accepted: April 16, 2007

© 2007, National Health Personnel Licensing Examination Board of the Republic of Korea

This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

  • 41,966 Views
  • 177 Download
  • 4 Crossref
  • After briefly reviewing theories of standard setting we analyzed the problems of the current cut scores. Then, we reported the results of need assessment on the standard setting among medical educators and psychometricians. Analyses of the standard setting methods of developed countries were reported as well. Based on these findings, we suggested the Bookmark and the modified Angoff methods as alternative methods for setting standard. Possible problems and challenges were discussed when these methods were applied to the National Medical Licensing Examination.
Recently the need for reliable and valid licensing examinations for health personnel is increasing in importance in Korea. This is due to its direct relation to the quality of health personnel and subsequently the nation’s health. The National Health Personnel Licensing Examination Board (NHPLEB) currently uses a cut score of 60-40% (which means 60% correct responses of overall tests and 40% correct responses of each subject) as a license requirement. These cut scores have been applied to various national examinations in Korea and are regarded as reasonable. From the perspective of psychometrics, however, it is not a valid way to set a cut score, especially for a licensing examination that is intended to discern those who acquire minimum competence from those who do not.
In this review, we first examined theories of standard setting and analyzed problems of the current cut scores. Then, we reported the results of need assessment on the standard setting among medical educators and psychometricians. Analyses of the standard setting methods of developed countries were reported as well. Based on these findings, we suggested new methods of standard setting and discussed possible problems and challenges when applied to National Medical Licensing Examination (NMLE).
A cut score indicates the minimum level of knowledge or skill that a candidate must have acquired to perform health professions. Cut scores are professionally expressed value on candidates’ competency, which reflects educational philosophy and consensus among those related. This means that setting cut scores should be understood not as a mathematical technique but as a complex policy-making process.
Setting cut scores usually progresses as follows: 1) Deciding on the type of standard (absolute vs. relative standards), 2) Deciding on the method for setting standards, 3) Selecting the judges, 4) Holding the standard setting meeting, 5) Calculating the standard, 6) Checking the results [1].
There are several methods to set standards. Generally accepted methods are as follows; contrasting groups method, borderline group method, Nedelsky’s method, Angoff’s method, Ebel’s method, and so on (for details, refer to [2]). Among these, the Angoff method had been the most popular method used for multiple-choice tests by the 1990s [3]. However, this method has a fundamental drawback, which is that it is practically impossible to estimate the probability that a minimally acceptable person would answer each item correctly. To accommodate this drawback, several modifications were suggested, one of which will be described in detail later in this paper.
The current 60-40% system was arbitrarily set under Japanese colonial rule, with no educational and philosophical basis. In addition, it can result in gross misclassification, such as the unfortunate failure of the competent and the fortunate pass of the non-competent depending on test difficulty.
Table 1 clearly demonstrates this phenomenon. In Table 1, the pass rate is approximately 95%. However, the pass rates of 2001 and 2003 drop to approximately 85%, lower than other years. Though not shown in Table 1, the pass rate of 1995 and 1996 was about 70%, a rate that brought mass confusion among examinees.
In principle, the difficulty of criterion-referenced test such as licensing examination is not adjusted before examination in consideration of optimal passing rate. In Korea, it is not practically possible to adjust the degree of difficulty in advance due to the drainage of test items. Instead, examiners are expected to make good questions asking basic knowledge and skill that are required to perform health professions. Since the current 60-40% criterion does not reflect this minimum level of competence, it should be changed.
Surveys were conducted to gather input from psychometricians, medical educators, and examiners. There were 38 respondents for the survey, and it was carried out from Januray 7th to 12th in 2005. We asked whether and why they thought the current cut score was valid and their suggestions for improvement. The results are in Table 2.
The followings are suggestions for improvement. First, most of the respondents believed that absolute evaluation should be retained. To do so, transformed scores can be a viable alternative. Second, examination objectives should be developed. Third, investment in faculty training for the item development is required. Fourth, scientific and systematic methodology is needed to analyze test items, to control the degree of difficulty. Fifth, item bank should hold more realistic items. Sixth, there should be flexibility in setting the cut score. For example, the Angoff method provides such flexibility. Finally, the cut score should be set under the agreement among stakeholders such as NHPLEB, medical representatives, and so on.
We analyzed the standard setting methods of several developed countries. The results are summarized in Table 3. From the table, we can see that the basic principles of standard setting are well maintained. That is, they stick to absolute standards, and put much effort into keeping the process of standard setting rational and reasonable. These results suggest Korea should adopt the standard setting method based on scientific methods of psychometrics.
We reached the agreement that Bookmark [4] and modified Angoff methods are appropriate considering the current situation in Korea. We intended to compare the two methods by applying them to real tests to increase the accuracy and validity of the results of the present study. However, we encountered a few realistic problems in carrying out such an empirical study. First, as a rule the NHPLEB does not release previous test items. Second, when releasing test items, the NHPLEB requested the research be carried out in the presence of NHPLEB personnel, which was not possible in terms of the time and resources allocated to this research. We therefore did not carry out an empirical study but instead described the process of the two methods. Both methods were recently applied to National Assessment of Educational Achievement in Korea.
The bookmark method
In 2002, the Bookmark standard setting was applied to discern those who achieved basic competency from those who did not among 3rd year elementary school students. It was the first application of the Bookmark method in Korea. The method allows a standard setting panel to be accustomed to the characteristics of the test and provides the panel with the results of standard application. The panel consists of subject experts. For example, the panel for setting national basic performance level in reading consists of professors of Korean language education, experienced teachers and researchers, administrators, and parents. The Bookmark method proceeds as follows [5].
  1. The panel received an ordered item booklet (OIB), which lists items in order of difficulty, from the easiest to the hardest. To make the OIB, item response theory (IRT) was applied and 2/3 probability of correct response was used as a scale score to represent examinee ability.

  2. The panel was divided into small groups-typically groups of six to eight people. In small groups, the panel examined each item in the OIB and discussed knowledge and skills required to answer the item correctly.

  3. After the group discussion, each panelist determined a cut score by placing a bookmark in the OIB based on his or her own judgment of what students with basic performance level should know and be able to do. The scale score of the marked item was the first demarcation.

  4. The panel engaged in the small group discussion again to compromise differences in their opinions. The panelists then marked their choice on the OIB, which was the second demarcation.

  5. Two small groups were combined into a mid-sized group to prevent one specific individual from dominating the discussion in small groups. After the group discussion, the panel was given another opportunity to change their choice and placed the bookmark on the OIB.

  6. Finally, all panelists gathered in one place and discussed their choice of bookmarks. After the discussion, they were given a final opportunity to change their choice. They then placed the final bookmark on the OIB.

  7. All the bookmark placements in the final round were gathered and the median was calculated to set the panel’s recommended cut score.

  8. Based on the cut score, the panel examined the items before the bookmark and wrote performance-level descriptors that represent a summary of the knowledge, skills, and abilities that students with the basic performance level must be able to demonstrate.

The bookmark method compensates for the drawback of the Angoff’s method, which is that the panel cannot correctly estimate item difficulty. In addition, people with little psychometric knowledge can understand the meaning and implication of the cut score because the bookmark method provides performance-level descriptors. Considering these theoretical merits, combined with the practical application of the method to the diagnostic assessment of the 3rd year elementary students, the bookmark method can be a viable alternative in setting the standard of national medical licensing examination.
Modified angoff
In the original Angoff method, there was no details of how to run an operational cut score study. Due to the lack of specificity in the original method, many modifications were suggested. The following is a general process of modified Angoff methods.
  1. The choice of a panel: Generally, the panel consists of subject experts, teachers, related administrators, and so on. In case of medical licensing examination, the panel may include doctors and professors of basic science and clinic, medical administrators, and so on. The panel should be content experts and know well the characteristics of examinees. There are diverse opinions on the appropriate number of the panel, but at least 10 panelists are required, while 15∼20 panelists are ideal [69].

  2. Achievement Level Description (ALD): Conceptualizing achievement level starts from policy definition. Government agency such as the Ministry of Education & Human Resources Development or the Ministry of Health & Welfare provides a policy definition of achievement levels. Based on this policy definition, the panel discusses and describes performance of each level. This description specifies what students at each level should know and be able to do in terms of knowledge, skills, and behaviors and provides an operational definition. Exemplary items can be provided as well. To save time, the facilitator may provide preliminary ALD with the panel, so that the panel starts to set the standard with some consensus on the characteristics of a borderline group.

  3. Practice: The panel practices with exemplary items. In case different types of items are mixed, they practice with each type of item. Especially for performance item, actual performance data should be provided, so that the panel can get a sense of the level of examinees.

  4. The first round of estimation: The panel is divided into small groups of three or four people. The panel solves actual items on the exam. After solving the exam, they check their answers with correct ones. Then, they estimate the probability of correct answers of a borderline group of examinees for each item. After individual estimation, the results are collected and the cut score is set at the sum of medians of each item. The cut score is posted and the result of cut score application is provided, so that the panel has opportunity to check whether it is realistic and change it if necessary.

  5. The second round of estimation: The panel is divided into mid-sized groups. Based on ADL and their experiences at the first round, they exchange their opinions. Then, they estimate the probability of correct answer of a borderline group of examinees for each item.

  6. The third round of estimation: All the panelists gather at one place and discuss. The focus is on the items showing large deviations. After the discussion, they make third estimations. The results are collected and posted. If the third round is enough, the cut point obtained at the third round becomes the final cut score. The cut score is transformed into a scale score.

  7. Description of minimum competency: If required, the panel writes what students at each level are able to do by analyzing items and students’ response to the items.

The modified Angoff method is widely used in foreign counties. The method has been applied to National Assessment of Educational Achievement since 2003 by Korea Institute of Curriculum & Evaluation (for details, see [10]). Since it is based on experts’ judgments and psychometric methods, the use of this method in setting the standard by the NHPLEB may enhance its expertise. In addition, the NHPLEB may be better prepared for any action taken by examinees who feel the cut score is unfair, because the method provides firm rationale behind the cut score.
Bookmark vs. Modified Angoff Methods
The two methods are compared in terms of process, time and cost, and advantages and disadvantages. The results are shown in Table 4. Comparing all these aspects, the Bookmark method is more recommended than the modified Angoff method theoretically. However, the modified Angoff method seems more realistic in Korea considering that IRT cannot be applied to the NMLE.
For the Bookmark and modified Angoff methods to be applied, there are a few practical issues to consider.
Need for committee: In a test-centered approach in setting the standards, there should be a committee for each subject. Since there are three subjects in NMLE, three committees are required. If twenty members are set for each committee, 60 members are in need, which would incur considerable costs. Insufficient budgeting, and consequent insufficient committee size, can lead to the loss of validity in setting the standard.
Need for psychometric analysis: Both methods recommended in the present paper are based on scientific analysis of psychometrics. Therefore, to apply the methods, classical item analysis and IRT should be applied in advance. For example, OIB in the Bookmark method requires pre-analysis based on IRT. The OIB costs extra time and cost, but it lessens the burden of the panel.
Need for extra cost: To apply a scientific and valid method, extra cost is inevitable.
No need for amendment of medical law: A question arises as to the need for amendment of medical law with a new standard setting method. However, the current law can be maintained even with a new method. One way to do so is to transform raw scores to scale scores. In Fig. 1, the cut score of raw scores is 250, which is transformed to 60. If amendment of the law can be done without much argument, changing the current law is another way to deal with legal issues.
Test equating method: In general, the cut score is not set every year. Instead, a statistical technique is applied to make different tests comparable. This is called test equating [11, 12]. If tests are equated, the same cut score can be used continuously. However, it is not practically possible in Korea to apply test equating methods due to a drainage of test items. In this case, where the cut score is set every year, comparison of cut scores across different years would be reasonable.
In the present paper, we attempted to suggest a valid and reasonable way to set the cut score of the NMLE. Based on the review of various theories, the need assessment, and the results of the analyses of foreign countries’ systems, we suggested the Bookmark and the modified Angoff methods as viable alternatives to the current system. To apply these new methods, several issues must be resolved beforehand. First, the philosophical meaning of licensing examination itself should be reconsidered. Under the current fixed cut score, test difficulty should be adjusted in advance in order to prevent a radical change in pass rates, which violates the concept of licensing or certifying examinations.
Second, philosophical reconsideration of the standard setting is required. The current 60-40% was set without any philosophical or educational rationale. Therefore, we need to set the cut score based on the meaning and philosophy of license examination.
Third, the issue of security and copyright of test items should be resolved. Currently, the NHPLEB as a rule does not release test items. However, examinees release the items to their peers after their examination, making it possible to almost restore the original test. This is obviously infringement of copyright. In Korea, however, this kind of act is not regarded as serious and those related show no moral remorse. This common practice of plagiarism should be controlled in the near future.
We hope our research facilitates the discussion of the standard setting in Korea, and contributes to produce competent medical professionals.
Support for this research was provided by National Health Personnel Licensing Examination Board of Korea.

This article is available from: http://jeehp.org/

Fig. 1.
An example of the application of the current 60% cut score.
jeehp-4-1-g001.gif
Table 1.
Pass Rate of Korean National Medical Licensing Examination
Year Examinees (N) Examinees who passed the exam (N) Pass rate (%)
2005 3,618 3,372 93.2
2004 3,881 3,760 96.9
2003 3,647 3,159 86.6
2002 3,578 3,314 92.6
2001 3,262 2,796 85.7
2000 2,961 2,772 93.6
1999 3,091 2,871 92.9
Table 2.
Results of need assessment (N=38)
Do you think the current cut score is valid? And why?
Yes (N=22) No (N=16)
  • - Criterion-referenced evaluation such as the current cut score is appropriate because the test is a licensing examination.

  • - Since almost all of the licensing examinations in Korea adopt the 60-40% cut score, it does not seem problematic.

  • - Even though the current cut score is not valid in a theoretical sense, there is no alternative, considering the cost of setting standards and the drainage of test items.

  • - An absolute cut score is not valid in the case of perennial fluctuation of test difficulty, such as in Korea.

  • - The current test items are not appropriate to evaluate basic medical knowledge.

  • - The reliability and validity of the test should be checked and skills as well as knowledge should be evaluated.

  • - The current cut score does not provide a social function to control quality of doctors.

Table 3.
Standard setting methods of the developed countries
Nation Test Standard setting methods
Australia Australian Medical Council (AMC) examination for overseas trained medical practitioners No examination for citizens Transformed scores
  Separate scoring of general items (200) and core items (60)
  General items 250/500
  Core items 300/500
Canada Medical Council of Canada Qualifying Examination (MCCQE) I &II Part I: Nedelsky Method
Part II: Angoff Method
OSCE: Boderline Group Method
England Professional and Linguistic Assessments Board (PLAB) Written: Angoff Method
Skill: Boderline Group Method (under consideration)
Ireland Temporary Registration Assessment Scheme (TRAS) Part I (MCQ) Equalized score of 45%; Penalty count system
Part II (OSCE): 85–90%
New Zealand New Zealand Registration Examination (NZREX) Written: Modified Angoff
Skill: Contrasting Groups Methods
U.S.A. United States Medical Licensing Examination (USMLE) Modified Angoff
U.S.A. Comprehensive Osteopathic Medical Licensing Examination (COMLEX) A criteria-referenced method is used and pass rate is about 90%.
Common features Criteria-referenced methods are used.
No country uses a fixed cut score applied to original scores.

OSCE: Objective Structured Clinical Examination; MCQ: Multiple Choice Questions.

Table 4.
Comparison of the bookmark and modified Angoff methods
Bookmark Modified Angoff
Process The panel places a bookmark on the hardest item to a borderline group on OIB. The panel estimates the probability of correct answer of a borderline group for each item.
Time In a day At least two days
Cost Due to little burden on the panel, costs are comparatively low. Pre-analysis by psychometricians is required, but there is no need for OIB.
But, an additional cost for OIB should be set. When the exam contains many items, the whole process takes much time and this increases costs.
Advantages Shorter time and lower cost since it was already applied in Korea, its validity was tested. Since it is a classic method in psychometrics, it is easy to understand and explain.
Preparation is relatively small.
Disadvantages Preparation such as OIB is required. Since estimation should be made for each item, the panel has a lot of work to do.

OIB: Ordered Item Booklet.

  • 1. Norcini JJ. Setting standards on educational tests. Med Educ 2003;37:464-9. ArticlePubMed
  • 2. Zieky MJ. So much has changed: How the setting of cutscores has evolved since the 1980s. In: Cizek GJ, editor. Setting performance standards. Mahwah (NJ): Lawrence Erlbaum Assoociates; 2001. p. 19-52.
  • 3. Mehrens WA. Methodological issues in standard setting for educational exams. In: Crocker L, Zieky M, editors. Proceedings of Joint Conference on Standard Setting for Large-Scale Assessments; 1994 Oct 5–7; Washington, DC: Washington, DC: U.S. Government Printing Office; c2005. 221-63.
  • 4. Lewis DM, Mitzel HC, Green DR. Standard setting: a bookmark approach. Paper presented at: IRT-based standard setting procedures utilizing behavioral anchoring. The 1996 Council of Chief State School Officers National Conference on Large-Scale Assessment. 1996 June; Pheonix, AZ:
  • 5. Lee G. A psychometric approach to setting a passing score on Korean National Medical Licensing Examination. J Educ Eval Health Prof 2004;1:5-14. ArticlePDF
  • 6. Mehrens WA, Popham WJ. How to evaluate the legal defensibility of high-stakes tests. Appl Meas Educ 1992;5:265-83. Article
  • 7. Cizek GJ. An NCME instructional module on: Setting passing scores. Educa Meas: Issues Pract 1996;15:20-31. Article
  • 8. Norcini JJ, Shea JA. The credibility and comparability of standards. Appl Meas Educ 1997;10:39-59. Article
  • 9. Zieky MJ, Livingston SA. Manual for setting standards on the basic skills assessment tests. Princeton (NJ): Educational Testing Service; 1977.
  • 10. Bahn JC. Scale and test equating of National Assessment of Educational Achievement. Paper presented at: 2004 Annual Fall Meeting of the Korean Society for the Study of Education. 2004 October 29–30; Daejeon, KR.Korean:
  • 11. Kolen MJ, Brennan RL. Test equating: methods and practices. New York (NY): Springer-Verlag; 1995.
  • 12. Lee G, Lewis DM. A generalizability theory approach toward estimating standard errors of cut scores set using the Bookmark standard setting procedure. Paper presented at: 2001 annual meeting of the National Council on Measurement in Education. 2001 April 11–13; Seattle, WA:

Figure & Data

References

    Citations

    Citations to this article as recorded by  
    • Predicting medical graduates’ clinical performance using national competency examination results in Indonesia
      Prattama Santoso Utomo, Amandha Boy Timor Randita, Rilani Riskiyana, Felicia Kurniawan, Irwin Aras, Cholis Abrori, Gandes Retno Rahayu
      BMC Medical Education.2022;[Epub]     CrossRef
    • Possibility of independent use of the yes/no Angoff and Hofstee methods for the standard setting of the Korean Medical Licensing Examination written test: a descriptive study
      Do-Hwan Kim, Ye Ji Kang, Hoon-Ki Park
      Journal of Educational Evaluation for Health Professions.2022; 19: 33.     CrossRef
    • Applying the Bookmark method to medical education: Standard setting for an aseptic technique station
      Monica L. Lypson, Steven M. Downing, Larry D. Gruppen, Rachel Yudkowsky
      Medical Teacher.2013; 35(7): 581.     CrossRef
    • Standard Setting in Student Assessment: Is a Defensible Method Yet to Come?
      A Barman
      Annals of the Academy of Medicine, Singapore.2008; 37(11): 957.     CrossRef

    Figure
    We recommend

    JEEHP : Journal of Educational Evaluation for Health Professions