Introduction

With the development and changes in the dental care environment, dental hygienists need to have practical skills as well as the ability to cope with various clinical situations actively [1], and an adequate curriculum is required to cultivate the professional manpower demanded in the clinical environment. This requires provision for sufficient practical experience and time in the dental hygiene curriculum, and the number of preclinical practice hours must be increased to improve the clinical proficiency of dental hygiene students [2]. However, due to the current coronavirus disease 2019 (COVID-19) pandemic, opportunities for hands-on training on clinical sites and on campus are limited. To provide the best education in a limited environment, the quality of the practical training needs systematic management. Quality control of curriculum not only improves the curriculum and satisfies the needs of students but also improves the quality of vocational education to nurture talents with on-site practical skills [3]. In Korea, the Dental Hygiene Education Evaluation Preparatory Committee has been established for quality control of education and the development of evaluation indicators. Evaluation items for in-school clinical practice subjects suggested by the preparatory committee include implementation of a practical curriculum, cost of practical training per student, organization and the use of practice guidebook, placement of students in clinical practice, and autonomous training [4]. However, specific standards to evaluate the quality management in in-school clinical practical training are lacking.

In this study, a systematically structured CIPP evaluation model was used to establish in-school clinical practical training operation management evaluation criteria to expand the evaluation target and enable rational decision-making by decision-makers in operating the curriculum. The CIPP evaluation model comprises four stages: contextevaluation, input evaluation, process evaluation, and product evaluation. Context evaluation detects undesirable circumstances or unmet needs and diagnoses problems, while input evaluation is used to provide information on resource utilization for the successful achievement of the goals. Process evaluation is used to identify the shortcomings of program implementation and procedural measures and collect information for revising and supplementing the program procedures and implementation methods. Product evaluation measures and interprets the result at the end of the program [5,6].

Previous studies on evaluation criteria have assessed job satisfaction, infection, supplementary education, and on-site clinical practical training; however, studies on in-school clinical practical training are lacking. Therefore, the purpose of this study was to develop operational management evaluation criteria for in-school clinical practical training and provide basic data for quality management and standardization of clinical practice management.

Methods

1. Study participants

This study was approved by the Korean Public Institutional Review Board (IRB) (P01-202109-22-005).

An expert interview was conducted with two dental hygiene professors who had been in charge of the practical training for more than 20 years and one expert with industry experience. The Delphi survey was conducted with 30 dental hygiene professors who had been in charge of practical training for more than 5 years and dental hygienists working in clinics. The survey was conducted from October 1, 2021 to October 10, 2021 through e-mails. The study was conducted from October 20, 2021 to November 20, 2021 on 252 clinical dental hygienists and professors who had been in charge of practical training. Based on a previous finding that stated the requirement of five to 10 times greater number of patients than the total number of items [7] and a possibility of participant drop-out, a total of 252 participants were explained the purpose of this study. The participants provided written consent for participation. The number of participants in this study was five times greater than the total number of items, and this study was conducted through face-to-face interviews and e-mails or using Naver forms. In the final analysis, data from 240 participants were included, excluding data of those who did not provide answers to certain items or provided insincere responses.

2. Procedures and methods

1) Evaluation criteria draft composition

The evaluation criteria draft was based on previous study findings on the CIPP evaluation model [8-11], training evaluation [12,13], dental hygiene clinical practice [14], and expert opinions. The evaluation area, items were classified according to the CIPP evaluation stage, and similar items were grouped into one. As a result, a total of 53 items were devised.

2) Validity verification

The content validity of the evaluation criteria draft was revised to a total of 44 items according to expert interviews. Then, the items were evaluated for content and suitability through the Delphi survey. Items that were lower than the cut-off value were deleted, and a total of 37 items remained. The main survey was conducted to assess the construct validity of the modified items. Exploratory factor analysis was conducted to construct a total of 36 items for the evaluation criteria.

3) Data analysis

SPSS 25.0 program (IBM SPSS Statistics, IBM, Corp.Armonk, NY, USA) was used for analyzing the collected data. Frequent analyses were conducted for the general characteristics of participants, and the mean, standard deviation, and content validity ratio (CVR) were calculated for data collected through the Delphi survey. Values with an average value below 0.4 and a CVR less than 3.3 were deleted after discussion with the experts [16]. The collected data were analyzed through exploratory factor analysis (EFA). The Kaiser-Meyer-Olkin (KMO) value and Barlett’s sphericity were calculated to check the suitability for EFA. Principal component analysis was conducted using Varimax rotation. The cut-off value for factor loading and commonality was 0.4 [17,18], and the consistency reliability of the items was measured using Cronbach’s α.

Results

1. General participant characteristics

The number of participants between the age of 30 and 39 in the Delphi survey was 40% and in the main survey was 31.7%. The ratio of professors and clinical dental hygienists was higher in the main survey (54.6%) than that in the Delphi survey. Approximately 33.3% and 31.7% of the participants in the Delphi and main surveys had experiences of 15 to 19 years, respectively. The highest number of participants (33.3%) in the Delphi survey lived in Seoul, while that (52.1%) in the main survey lived in Gyeonggido <Table 1>.

Table 1. General characteristics of the subject
SurveyCharacteristicsDivisionN (%)
Delphi surveyAge (yrs)30 - 39 12 (40.0)
40 - 49 10 (33.3)
50 - 59 7 (23.3)
≥ 60 1 (3.30)
DivisionClinical dental hygienist 15 (50.0)
Professor 15 (50.0)
Total career (yrs) 5 – 9 7 (23.3)
10 – 14 9 (30.0)
15 – 19 10 (33.3)
≥ 20 4 (13.3)
Area Seoul 10 (33.3)
Gyeonggi 8 (26.7)
Chungcheong 2 ( 6.7)
Jeolla 5 (16.5)
Gyeongsang 2 ( 6.7)
Gangwon 1 ( 3.3)
Main surveyAge (yrs) ≤ 29 41 (17.1)
30 – 39 76 (31.7)
40 – 49 63 (26.3)
50 – 59 53 (22.1)
≥ 60 7 ( 2.9)
Division Clinical dental hygienist 109 (45.4)
Professor 131 (54.6)
Total career (yrs) 1 – 4 44 (18.3)
5 – 9 56 (23.3)
10 – 14 34 (14.2)
15 – 19 76 (31.7)
≥ 20 30 (12.5)
Seoul 80 (33.3)
Gyeonggi 125 (52.1)
Chungcheong 19 ( 7.9)
Jeolla 4 ( 1.7)
Gyeongsang 6 ( 2.5)
Gangwon 2 ( 0.8)
Etc 4 ( 1.7)

2. Content validity verification

The evaluation criteria draft comprised 53 items in total, and the content validity of the items was assessed through expert interviews and the Delphi survey. Context evaluation comprised six items on demand analysis and two items on goal setting. Input evaluation comprised seven items on the operation plan and five items on the practical training support system. Process evaluation comprised five questions on practical training operation, nine items on practical training guidance, four items on practical training evaluation, and five items on practical training support. Product evaluation comprised two items on satisfaction and eight items on results and achievement.

1) Expert interview

To construct universally valid and objective evaluation criteria, three experts with more than 20 years of education and clinical experience were interviewed. As a result, a total of 44 items were constructed according to the discussions on the necessity of items, appropriateness of domain classification, added and deleted contents, and duplicated and revised contents <Table 2>.

Table 2. Draft evaluation criteria and expert opinion
Evaluation areaCategoryItemExpert opinion
ContextDemand analysis Job environment analysis‘Delete’ because it is difficult to reflect the opinions of students without clinical experience
Job analysis
Curriculum analysis
Survey on industrial needs‘Delete’ because the instructor’s opinion was reflected in other items
Survey on student needs
Research on the professor’s needs
Setting goalsSelection of goals for practice
Feasibility of the practice goals
InputPractice operation planSelection of training contents
Presenting a training guide method ‘Correction’ to selection of practice guidance method
Evaluation items and evaluation contents plan‘Correction’ to selection of evaluation items and evaluation contents
Writing a syllabus
Development of practice guidebook
Allocation of trainees ‘Delete’ because it overlaps with the appropriateness of the training staff
Guidance plan for students maladjusted to practice
Sharing practice instruction‘Additional’ because a large number of instructors are assigned to the class
Practice support planBudget management plan
Composition of training guidance personnel
Plan to expand practical equipment
Securing a practice room
Lab safety management guidance
ProcessPractice operationPreparation of practice equipment‘Delete’ overlaps with the expansion of practical equipment in the input evaluation part
Appropriateness of the practice environment
Appropriateness of practice contents
Appropriateness of the number of trainees
Appropriateness of practice time
ProcessTraining guidanceOrientation held
Suggestion of practice goals
Motivation
Inducing students’ interest and responsibility‘Delete’ because it overlaps with motivation
Support for students who are stuck in practice
Efficient mutual practice‘Correction’ by encouraging student cooperation
Communication‘Correction’ to practice - related communication
Convergence of practice - related requirements
Feedback on requirements
Practice evaluationSetting evaluation criteria according to practice goals‘Deleted’ because the evaluation criteria are also included in the evaluation items and evaluation contents
Presentation of practice evaluation criteria
Objectivity and relevance of practice evaluation criteria
Feedback of practice evaluation results
Training applicationManagement guidance on infection prevention
Appropriateness of the practice budget
Appropriateness of qualifications of practice instructors‘Deleted’ because the qualification criteria are met in the recruitment criteria when recruiting teachers
Appropriateness of practice place
Sharing and solving safety management problems ‘Correction’ by efforts to share and solve infection and safety management
ProductSatisfactionSatisfaction survey‘Delete’ because it overlaps with the achievement of the practice goals
Satisfaction results are reflected in the next practice plan and improvement
Practice results and achievementMeasurement of field adaptation
Measurement of whether the practice goals has been achieved
Measurement of improvement of major practical ability‘Delete’ because it overlaps with the achievement of the practice goals
Measurement of changes in attitudes toward dental hygienists’ jobs‘Delete’ as a point that is difficult to actually measure
Measurement of change in attitude toward participation in practice
Documentation of practice results
Share the results of the practice
Practice results are reflected in curriculum improvement and planning
2) The Delphi survey

The evaluation criteria revised through expert interviews were further evaluated using the Delphi survey. Items with an average value below 0.4 and a CVR below 3.3 were deleted after discussions with experts. As a result, six and one items were deleted in the first and second Delphi survey, respectively, and a total of 37 items remained.

In process evaluation, the item ‘likelihood of reaching the practical training goal (M 3.97, CVR 0.13)’ presented an average value of less than four and a CVR value of less than 3.3. According to the expert opinion, the item was similar to the ‘assessment of practical training goal achievement’; hence, it was deleted.

In the input evaluation, each of the items ‘guidance plan for students who have trouble adapting to practical training (M 3.87, CVR 0.13)’, ‘sharing practical guidance information (M 3.94, CVR 0.47)’, and ‘securing a practice room (M 3.87, CVR 0.13)’ presented an average value less than four and a CVR value less than 3.3. The item ‘guidance plan for students who have trouble adapting to practical training’ was similar to ‘support for students showing no improvement in practical training’, and the item ‘sharing practical guidance information’ was also assessed in ‘sharing practical training results’. Additionally, the item ‘securing a practice room’ was deleted as practical training was mostly provided after securing a practice room. Item ‘budge management plan (M 3.97, CVR 0.46)’ presented an average value of less than four; however, according to the expert opinion, the item was an essential element for practical training; hence, it was not deleted.

In the process evaluation, the item ‘appropriateness of practical training location (M 4.43, CVR 0.26)’ presented an average value above four, suggesting its relative importance. However, the item presented a CVR value of less than 3.3. According to the expert opinion, this item could be included in another item ‘appropriateness of practical training environment’; hence, the item was deleted. In product evaluation, items ‘assessment of changes in attitude toward practical training participation (M 3.42, CVR 0.20)’ and ‘documentation of practical training results (M 3.90, CVR 0.46)’ presented an average value of less than four and a CVR value of less than 3.3. As the item ‘assessment of changes in attitude toward practical training participation’ was difficult to measure because of the lack of an evaluation scale, which can clearly measure changes in attitude and time constraints, the item was deleted. In addition, the item ‘documentation of practical training results’ was deleted according to the expert opinion that the syllabus and teaching guidance plans included general information <Table 3>.

Table 3. Delphi research results of draft evaluation criteria
Evaluation areaDivisionItem1st2st
MSDCVRMSDCVR
ContextDemand nalysisQ1. Job environment analysis4.43 0.77 0.81 4.170.59 0.86
Q2. Job analysis4.47 0.81 0.73 4.63 0.49 1.00
Q3. Curriculum analysis4.13 0.90 0.46 4.230.720.66
Q4. Survey on industrial needs 4.00 0.90 0.42 4.00 0.93 0.41
Setting goalsQ5. Selection of goals for practice4.10 0.80 0.46 4.33 0.710.46
Q6. Feasibility of the practice goals3.97 0.74 0.13
InputPractice operation planQ7. Selection of practice contents4.47 0.77 0.66 4.53 0.50 1.00
Q8. selection of practice guidance method4.53 0.81 0.73 4.59 0.56 0.73
Q9. Selection of evaluation items and evaluation contents4.33 0.92 0.66 4.50 0.630.73
Q10. Writing a syllabus4.27 0.86 0.60 4.00 0.64 0.61
Q11. Development of practice guidebook4.23 0.93 0.46 4.07 0.74 0.46
Q12. Guidance plan for students maladjusted to practice4.10 0.62 0.46 3.87 0.76 0.13
Q13. Sharing practice instruction 3.94 0.78 0.47
Practice support planQ14. Budget management plan 4.37 0.71 0.66 3.97 0.71 0.46
Q15. Composition of training guidance personnel4.63 0.66 0.82 4.30 0.65 0.86
Q16. Plan to expand practical equipment4.47 0.62 0.86 4.33 0.71 0.80
Q17. Securing a practice room 3.87 0.87 0.13
Q18. Lab safety management guidance 4.50 0.67 0.82 4.43 0.62 0.73
ProcessPractice operationQ19. Appropriateness of the practice environment4.60 0.56 0.93 4.43 0.50 1.00
Q20. Appropriateness of practice content4.63 0.55 0.93 4.60 0.49 1.00
Q21. Appropriateness of the number of trainees4.40 0.77 0.82 4.23 0.67 0.73
Q22. Appropriateness of practice time4.47 0.68 0.81 4.47 0.68 0.80
Training guidanceQ23. Orientation held4.30 0.87 0.67 4.10 0.84 0.53
Q24. Suggestion of practice goals 4.30 0.91 0.53 4.37 0.66 0.86
Q25. Motivation4.50 0.77 0.85 4.47 0.62 0.86
Q26. Support for students who are stuck in practice4.60 0.56 0.93 4.63 0.68 0.93
Q27. Encouraging student cooperation 4.30 0.53 0.60 4.13 0.77 0.66
Q28. Practice-related communication4.77 0.50 0.93 4.57 0.67 0.81
Q29. Convergence of practice - related requirements4.47 0.50 1.00 4.27 0.64 0.86
Q30. Feedback on requirements 4.70 0.46 1.00 4.37 0.61 0.93
Practice evaluationQ31. Presentation of practice evaluation criteria4.40 0.77 0.66 4.33 0.66 0.80
Q32. Objectivity and relevance of practice evaluation criteria4.67 0.60 0.86 4.43 0.62 0.86
Q33. Feedback of the practice evaluation result 4.67 0.60 0.86 4.43 0.62 0.86
Training application Q34. Management guidance on infection prevention4.73 0.52 0.93 4.60 0.56 0.93
Q35. Appropriateness of the practice budget4.33 0.66 0.93 4.23 0.69 0.93
Q36. Appropriateness of the practice place 4.43 0.67 0.26
Q37. Efforts to share and solve infection and safety management4.67 0.61 0.86 4.37 0.76 0.66
ProductSatisfactionQ38. Satisfaction survey4.03 0.99 0.46 4.10 0.66 0.66
Q39. Satisfaction result is reflected in the next practice plan and improvement4.37 0.80 0.73 4.23 0.62 0.86
Practice results and achievementQ40. Measure whether the practice goals has been achieved4.20 0.71 0.81 4.13 0.770.66
Q41. Measurement of change in attitude toward participation in practice3,42 0.77 0.20
Q42. Documentation of practice results 3.90 0.92 0.46
Q43. Share the results of the practice4.07 0.86 0.68 4.03 0.71 0.53
Q44. Practice results are reflected in curriculum improvement and planning4.40 0.93 0.81 4.20 0.80 0.73

3. Construct validity verification

To determine the suitability of EFA, the KMO scale and Barlett’s sphericity were calculated. The KMO scale was 0.895, and the approximate chi-square value for Bartlett’s sphericity test was 5,181.213 (p<0.001), satisfying the conditions of factor analysis [19]. Factor analysis extracted seven factors with an eigenvalue greater than one and a cumulative explanatory rate of 64.413%.

Factor Ⅰ was grouped into eight items with an explained variance of 11.416%. Factor Ⅰ was named “practical training operation” as it comprised the following items on practical training operation: ‘composition of practical guidance manpower’, ‘plan for expanding training equipment’, ‘appropriateness of practical training environment’, ‘budget operation plan’, ‘appropriateness of trainees’, ‘appropriateness of training contents’, ‘appropriateness of operating budget’, and ‘appropriateness of practical training time’.

Factor Ⅱ was grouped into six items with an explained variance of 11.414%. Factor Ⅱ was named “communication and interaction” as it comprised the following items of communication and interaction between professors and students: ‘inducing student cooperation’, ‘supporting students stuck in practical training’, ‘motivating students’, ‘feedback on demands’, and ‘practical training-related communication’.

Factor Ⅲ was grouped into five items with an explained variance of 9.767%. Factor Ⅲ was named “practical training results and satisfaction” as it comprised the following items: ‘satisfaction survey’, ‘reflection of satisfaction results in next practical training plan and improvement’, ‘reflection of practical training results in educational improvement and plan establishment’, ‘assessment of goal achievement’, and ‘sharing practical training results’.

Factor Ⅳ was grouped into four items with an explained variance of 8.570%. Factor Ⅳ was named “practical training evaluation” as it comprised the following items: ‘presentation of practical training evaluation standards’, ‘objectivity and appropriateness of practical training evaluation standards’, ‘feedback of practical training evaluation results’, and ‘selection of evaluation items and contents’.

Factor Ⅴ was grouped into seven items with an explained variance of 8.496%. Factor Ⅴ was named “operation plan” as it comprised the following items: ‘orientation’, ‘practical training instruction method selection’, ‘practical training goal presentation’, ‘development of practical training guide’, ‘setting practical training goal’, ‘preparation of lesson plan’, and ‘selection of practical training content’.

Factor Ⅵ was grouped into four items with an explained variance of 8.330%. Factor Ⅵ was named‘demand analysis’ as it comprised the following items on demand investigation and analysis for setting goals in designing the education process: ‘analysis of work environment’, ‘task analysis’, ‘education process analysis’, and ‘industrial needs survey’.

Factor Ⅶ was grouped into three items with an explained variance of 6.420%. Factor Ⅶ was named‘infection and safety management’ as it comprised the following items on safety accidents: ‘infection prevention and control guidance’, ‘laboratory safety management guidance’, and ‘efforts to share and solve infections and safety management’ <Table 4>.

Table 4. Exploratory factor analysis results of evaluation criteria
ItemFactor ⅠFactor ⅡFactor ⅢFactor ⅣFactor ⅤFactor ⅥFactor ⅦCommunality
Q150.684 0.135 0.205 0.163 0.208 0.019 0.086 0.606
Q160.667 0.174 0.067 0.089 0.253 0.065 0.029 0.556
Q190.651 0.188 0.057 0.248 - 0.029 0.166 0.279 0.630
Q140.643 0.065 0.305 0.142 0.236 0.197 - 0.117 0.639
Q210.609 0.215 0.051 - 0.040 0.036 0.204 0.215 0.511
Q200.496 0.156 0.071 0.495 0.058 0.142 0.194 0.582
Q350.401 0.398 0.391 0.109 - 0.004 0.272 0.095 0.553
Q220.290 0.128 0.057 - 0.011 0.010 - 0.029 - 0.091 0.168
Q270.181 0.747 0.031 0.004 0.103 0.100 0.310 0.709
Q260.251 0.677 0.118 0.212 0.121 0.077 0.281 0.680
Q290.230 0.659 0.425 - 0.030 0.045 0.084 - 0.070 0.682
Q25- 0.054 0.657 0.045 0.181 0.439 0.072 0.038 0.669
Q300.288 0.605 0.346 0.257 0.131 0.010 0.107 0.664
Q280.194 0.604 0.163 0.430 0.079 0.004 0.074 0.625
Q380.086 0.141 0.816 0.070 0.098 0.074 0.119 0.727
Q390.160 0.120 0.770 0.157- 0.074 0.129 0.2510.742
Q440.106 0.227 0.548 0.367 0.284 0.154 0.213 0.647
Q400.710 0.247 0.486 0.396 0.254 0.386 0.116 0.537
Q430.182 0.217 0.463 0.365 0.246 0.106 0.080 0.507
Q310.023 0.175 0.219 0.719 0.3440.160 - 0.007 0.741
Q320.258 0.064 0.151 0.6950.225 0.256 0.208 0.736
Q330.197 0.359 0.185 0.616 - 0.066 0.328 0.031 0.694
Q90.160 0.115 0.371 0.452 0.291 0.101 - 0.418 0.650
Q230.140 0.135 0.072 0.107 0.713 0.092 0.301 0.662
Q80.355 0.053 - 0.072 0.256 0.576 0.230 0.166 0.611
Q240.045 0.359 0.438 0.233 0.555 0.034 0.004 0.686
Q110.382 0.243 0.164 0.078 0.536 0.220 - 0.004 0.574
Q50.133 0.415 0.470 0.044 0.526 0.048 - 0.041 0.694
Q100.283 0.176 0.408 0.036 0.458 0.300 - 0.306 0.673
Q70.367 - 0.015 - 0.040 0.333 0.4120.407 0.053 0.583
Q10.069 0.057 0.126 0.061 0.009 0.831 0.045 0.720
Q20.171 0.012 0.030 0.259 0.157 0.729 0.158 0.678
Q30.127 0.209 0.233 0.287 0.218 0.565 - 0.084 0.571
Q40.331 0.097 0.139 0.077 0.362 0.538 0.044 0.567
Q340.017 0.276 0.193 0.181 0.105 0.076 0.748 0.722
Q180.290 0.098 0.152 0.055 0.135 0.010 0.712 0.646
Q370.212 0.400 0.388 0.014 0.137 0.230 0.487 0.664
Eigen value3.767 3.766 3.223 2.828 2.804 2.749 2.119
Ratio of total variance explained (%)11.416 11.414 9.767 8.570 8.496 8.330 6.420
Cumulative ratio (%)11.416 22.829 32.596 41.167 49.663 57.993 64.413

FactorⅠ: practice operation; FactorⅡ: communication and interaction; Factor III: practice results and satisfaction; Factor Ⅳ: practice evaluation, FactorⅤ: practice operation plan; Factor Ⅵ: demand analysis; Factor Ⅶ: infection and safety management

4. The final evaluation criteria

Factor analyses showed that the item ‘appropriateness of practical training time’ in factor Ⅰ had a factor loading and commonality less than 0.4; as a result, the item was deleted. Other items showed inconsistency in meanings unlike in the statistical results; hence, they were removed. Items ‘composition of practical training guidance manpower’, ‘plan for expanding training equipment’, and ‘budget operation plan’ were moved from group Ⅰ to group Ⅴ ‘operation plan’ Items ‘orientation’ and ‘practical training goal presentation’ were moved from group Ⅴ to group Ⅰ ‘practical training operation’. Thus, the final evaluation criteria comprised 36 items in seven factors, including four items on demand analysis, eight items on operation plan, six items on practical training operation, six items on communication and interaction, three items on infection and safety management, four items on practical training evaluation, and five items on practical training results and satisfaction <Table 5>.

Table 5. Final evaluation criteria
FactorItemCronbach’s α
Demand analysisJob environment analysis0.760
Job analysis
Survey on industrial needs
Curriculum analysis
Practice operation planSelection of goals for practice0.757
Selection of practice contents
selection of practice guidance method
Development of practice guidebook
Composition of training guidance personnel
Budget management plan
Plan to expand practical equipment
Writing a syllabus
Practice operationOrientation held0.732
Suggestion of practice goals
Appropriateness of practice contents
Appropriateness of the number of trainees
Appropriateness of the practice environment
Appropriateness of the practice budget
Communication and interactionMotivation0.855
Encouraging student cooperation
Support for students who are stuck in practice
Practice - related communication
Convergence of practice - related requirements
Feedback on requirements
Infection and safety managementManagement guidance on infection prevention0.753
Lab safety management guidance
Efforts to share and solve infection and safety management
Practice evaluationSelection of evaluation items and evaluation contents0.812
Presentation of practice evaluation criteria
Objectivity and relevance of practice evaluation criteria
Feedback of the practice evaluation result
Practice results and satisfactionMeasure whether the practice goals has been achieved0.826
Satisfaction survey
Satisfaction result is reflected in the next practice plan and improvement
Share the results of the practice
Practice results are reflected in curriculum improvement and planning

5. Reliability of variable measurements

Cronbach's α for the developed seven factors were as follows: 0.760 for demand analysis, 0.757 for operation plan, 0.732 for practical training operation, 0.855 for communication and interaction, 0.753 for infection and safety management, 0.812 for practical training evaluation, and 0.826 for practical training results and satisfaction. Cronbach’s α for the total domain was 0.914. Cronbach’s α was greater than 0.7 for all factors, suggesting that the criteria were reliable [20] <Table 5>.

Discussion

In this study, we developed evaluation criteria and assessed their validity and reliability using literature review, expert review, and the Delphi and main surveys for the efficient operation and management of in-school clinical practice. The sub-factors of evaluation criteria were demand analysis, operation plan, practical training operation, communication and interaction, infection and safety management, practical training evaluation, and practical training results and satisfaction, including a total of 36 items.

Demand analysis comprised items basic for the establishment of a curriculum on vocational education. The changes in the dental clinical environment must be assessed to provide opportunities for realistic, practical training in the limited in-school clinical practice environment. Moreover, the duties and curriculum of dental hygienists must be analyzed to select the contents of practical training. Job analysis of dental hygienists ensures a guideline for selecting the contents and providing relevant training in the education environment. Taichman et al.[21] have emphasized the need for a job training program through job analysis to improve the profession, such as skills to cope with the rapidly changing environment, interaction with patients and other health care providers, and job improvement.

The operation plan comprised items on the plan and support system for the operation of the practical training. The selection of practical training goals is an essential element that must be conducted in the planning of the curriculum and is selected based on the analysis of the job environment and needs. After selecting the goals, realistic and standardized practical contents and methods of instruction are selected based on the proficiency and performance of the students. Then, a practical training guideline is developed. Additionally, through proper allocation of budget, practical training cost is secured, and a plan to expand the training equipment is established for adequate operation of the training system.

Practical training operation comprised items required for the implementation of the practical training program. In the first stage of the practical training, the goals of the program must be presented in detail through a comprehensive assessment of the curriculum characteristics, overall progress, and practical training based on the learning outcomes and core dental hygiene competencies [4]. As there may be individual differences in the performance of students, the overall performance level of students must be understood to determine the appropriateness of the contents and the number of trainees and accordingly modify the next training program. The practice room environment affects not only in-school but also on-site clinical practices. Shim et al.[22] have shown that students who were highly satisfied with the in-school clinical practice environment were highly satisfied with the on-site clinical experience. An appropriate operation budget is an essential element of practical training. In dental hygiene evaluation and certification standards, 80% or more of the total practical training budget must be used. More than 200,000 won must be secured for each student, and a detailed list and supply of required equipment and tools must be presented [4].

Communication and interaction comprised items related to exchanges among students and professors or among students during the practical training. Motivation and communication to induce active participation in training activities are important factors in determining the effects of practical training. Attitudes, such as learning motivation, enthusiasm, values, and decision, can influence learning, and learners with an active learning attitude can participate in learning in a motivated and efficient way to increase learning efficiency[23]. In-school clinical practice is mostly conducted in a limited time by a single instructor who shows a demonstration of skills directly or through mutual practice to a large number of students without on-site clinical experiences. In the course of practical training, students may find difficulties in the program, maladapt, and face problematic situations in mutual practice. Thus, based on individual differences of students, customized practical training that allocates time and autonomous practice outside class hours must be provided. In a previous study, Kim [24] has reported that developmental learning outcomes are observed when an appropriate teaching/learning method is provided according to the learning type during interviews with and coaching for trainees. Thus, Kim suggested the importance of developing suitable learning types and customized teaching and learning methods.

Infection and safety management comprised items on infection and injuries that may occur during practical training and safety management of training equipment. Current infection and safety management in practical training includes mandatory attendance for the training, provision of first-aid supplies while preparing for possible accidents during training, connecting with the health center in cases of potential accidents, and insurance coverage. Students participating in practical training lack the understanding of safety accidents and infections. Therefore, proper management and guidance must be provided to the students for their safety, and the occurrence of various problematic incidents during training must be shared to seek countermeasures.

Practical training evaluation comprised items that evaluated the performance of students during training. Evaluation of the practical training performance may be a sensitive interest for students. Thus, it is essential to select appropriate and objective evaluation items, contents, and standards that meet the goals of training and provide a clear explanation to the students. After completing the evaluation, feedback on the results must be assessed by the students for an opportunity to self-assess and improve their performance.

Practical training results and satisfaction comprised items that were implemented at the end of the training. A satisfaction survey of trainees is a tool to evaluate the effectiveness of the practical training curriculum. It acts as an important factor that can modify or improve the direction of the program. In addition, sharing and discussing the overall contents and the results of practical training with the department members can improve the quality of the program through step-by-step improvements and supplementation of limitations for the establishment of the next plans.

This study is meaningful because it developed its own evaluation criteria for the operation and management of in-school clinical practice. However, due to regional bias and a limited number of selected participants, the findings of this study cannot be generalized. To obtain generalized results in the future, larger sample size is necessary. In addition, it would be important to derive specific evaluation indicators and scales appropriate for evaluation criteria to verify the effectiveness and supplement the problems of evaluation.

Conclusions

This study was conducted to evaluate the quality control evaluation criteria of the in-school practical training course.

1. The evaluation criteria draft comprising 53 preliminary items was developed through a literature review.

2. The content validity of the items was verified through expert interviews and Delphi surveys. Based on the results of expert interviews, the number of items was reduced to 44. Following Delphi surveys, a total of 37 items were constructed.

3. The construct validity was verified through EFA in the main survey. As a result, the final criteria comprised 36 items, including four items on demand analysis, eight items on operation plan, six items on practical training operation, six items on communication and interaction, three items on infection and safety management, four items on practical training evaluation, and five items on practical training results and satisfaction.

4. Cronbach’s α was 0.760 for demand analysis, 0.757 for operation plan, 0.732 for practical training operation, 0.855 for communication and interaction, 0.753 for infection and safety management, 0.812 for practical training evaluation, and 0.826 for practical training results and satisfaction, and all values were satisfactory.

Based on these results, the developed evaluation criteria for in-school clinical practice operation and management presented adequate validity and reliability. Future studies with specific evaluation indicators and scales need to be developed for the quality management of the in-school clinical practice operation.