Literature DB >> 34084829

Eliciting preferences of professors and medical group students for evaluation methods of theoretical courses: An application of discrete choice experiment analysis.

Ali Kazemi Karyan1, Satar Rezaei1, Shokooh Etesami2, Leyla Pezhman2, Behzad Karami Matin1, Sajad Delavari3.   

Abstract

BACKGROUND: Evaluation has become an inseparable part of education process which gives feedback to students and professors to improve education quality. This study aimed to elicit preferences of professors and students about attributes of evaluation methods in theoretical courses in Kermanshah University of Medical Sciences, Iran, in 2018.
MATERIALS AND METHODS: Discrete choice experiment (DCE) method used for eliciting preferences of participants of the study. A narrative literature review and interview with eight professors and ten students conducted to determine attributes and levels of evaluation methods in the university. Furthermore, experimental design used for making final choice sets of the evaluation methods. We included 213 students and 30 professors in the study. Conditional logistic regression model performed to data analysis.
RESULTS: Most of the professors (36.67%) preferred to allocate up to 30% of evolution scores to midterm examination. However, the most percentage of students (30.45%) were agree to include midterm examination up to 15% of total scores. The majority of students prefer to examination questions compromise just presented materials, while 70% of professors prefer to include additional texts for evaluation examinations. In case of quiz examination, professors in comparison with students prefer that quiz should have higher proportion of total scores. DCE analysis indicated that professors and students preferred a mix of questions in examinations. In addition, additional resources beyond what is taught in class made utility for professors and disutility for students. Quiz, also, increased the utility of an evaluation package in professors.
CONCLUSION: The findings showed that there is a gap between preferences of professors and students regarding some attributes of evaluation methods such as student's discipline, examination materials, and quiz. Further studies are needed to examining other attributes of evaluation methods in theatrical and practical courses in Iran and other contexts. Copyright:
© 2021 Journal of Education and Health Promotion.

Entities:  

Keywords:  Discrete choice experiment; evaluation; preferences; professors; students; theoretical courses

Year:  2021        PMID: 34084829      PMCID: PMC8057175          DOI: 10.4103/jehp.jehp_540_20

Source DB:  PubMed          Journal:  J Educ Health Promot        ISSN: 2277-9531


Introduction

Education is one of the important factors in sustainable development of countries.[123]. It could improve economic growth, equity, and welfare.[345] Thus, educational institutions are trying to transfer knowledge and skills to the population to improve economic development of nations.[6] The evaluation has become an inseparable part of education which gives feedback to students and teacher to improve education process.[7] This could be a continual process during education or a point at the end. The results of evaluation could be very useful in designing other education or redesigning current courses to improve quality.[8] Thus, it could improve education process and reduce deficiencies. On the other hand, it could rank students which encourage them for more effort and let other agents to know students' learning and expertise. In medical education, evaluation is more important than other fields. Medical students should learn their required knowledge, skills, and expertise with the minimum error to practice them for patients.[9] Professors need to evaluate their students and rank them based on their learning and skills. Thus, they try to do several types of evaluation methods to reach a reliable conclusion. However, different professors use different methods and as a result they reach to different outcomes which may not be reliable.[10] In fact, education systems need valid and reliable evaluation methods to evaluate university outcomes. If student evaluation be similar in universities, health systems could rely more on students' scores in universities and could easily compare their knowledge.[9] However, there is a plenty of research on students' evaluation method, there is not enough literature on preparing a valid model for student and learning evaluation at universities. Most of the researches are done in descriptive form and tried to present the methods which are used by universities' professors and lecturers for student evaluation. Some others such as Pearson et al. (2012), tried to find evaluation criteria in scoring students.[11] Other researches use techniques such as analytical hierarchy process,[1213] neural network,[14] fuzzy logic[15] or data mining[16] to evaluate education. Most of the mentioned models and methods are not applicable in practice and mainly are not based on preferences of instructors and students. Thus, it seems that these researches could not fulfill requirements of education system. Considering preferences of students in different aspects of medical education can lead to improving the quality of health-care services that they will provide in the future.[1718] Discrete choice experiment (DCE) could present models for analyzing preferences which have previously used in education valuation. Kennelly et al. (2015) found that business students prefer to have online assignments instead of paper assignments.[19] In addition, another study at McMaster University used DCE to investigate preferences of medical students about problem-based learning. This study indicated that students preferred small tutorial groups, and programs that have web-based support.[20] An increasing number of studies in Iran investigates learning styles and evaluation methods among students and faculty members.[21222324] However, the evidence is rare about preferences of medical students and faculty members about attributes of current evaluation methods in Iran. This evidence could help education planners to analyze students and professors' preferences in scoring and evaluation. In this way, we could find evaluation criteria and their importance in scoring from different viewpoints. Thus, in the present research, we use DCE approach to find professors and students preferences about attributes of evaluation methods in theoretical courses in Kermanshah University of Medical Sciences (KUMS) in the west of Iran. The results could help educational planner to select a logical and valid attributes for evaluations.

Materials and Methods

This study was a descriptive-analytical study that used DCE method of analysis to elicit preferences of professors and students of KUMS about attributes of evaluation methods in theoretical courses during 2018. For this purpose, following steps of DCE method were done:

Determining attributes and levels

The basic hypothesis of DCE method is that we can describe the goods, services, methods, etc., using their attributes. It is suggested to use literature review and qualitative studies such as interview with experts for determining attributes and attributes-levels.[2526] Therefore, we applied a narrative literature review and interview with eight professors and 10 students to determine attributes and levels of evaluation methods in the university. The experience of at least 5 years of teaching in the university was inclusion criteria for professors to include in qualitative study. In addition, students passed at least 20 theoretical courses in the time of study. All professors had academic degrees related to medical sciences such as health-care management, health education, health economics, nursing, and medicine. We continued interviews until achieving saturation in information. After reviewing literature and the interviews, the research team determines six main attributes to include in final design. The attributes and levels that included in final design are shown in Appendix 1.
Appendix 1

Attributes and attributes-levels of evaluation plans of students in theoretical courses in Iran

RowAttributesLevels
1Student’s discipline1. Not included (0%), 2. up to 5% of total score, 3. up to 10% of total score
2Midterm exam1. Not included, 2. up to 15% of total score, 3.up to 30% of total score, 4.up to 40% of total score
3Questions1. Essay, 2. Multiple choice, 3. A mix of essay and multiple choice questions, 4. a mix of essay, multiple choice, fill in the blanks, yes-no, true-false questions
4Materials of examination1. Presented material, 2. Presented materials and additional references.
5Final examination1. Up to 50% of total score, 2.up to 70% of total score, 3. up to 90% of total score
6Quiz1. Not included (0%), 2. up to 5% of total score, 3. up to 10% of total score

Experimental design

By defining attributes and levels, scenarios were created using a combination of attributes and levels. In other words, this step of the study involves deciding which choices to make to present to the study participants.[27] The output of experimental design is set of choice sets that use to eliciting the preferences of participants. A choice set consists of the choices that the individual chooses. There are several ways to design the choice sets of study. However, in many cases, respondents are offered two or more options to choose one among them. In this study, 7 choice sets, each with two scenarios, made by D-efficiency criteria to optimize the final design. Appendix 2 shows a choice set included in the study.
Appendix 2

One of the choice set included in the study

AttributesPlan APlan B
Student’s disciplineUp to 5% of total scoreNot included (0%)
Midterm examinationUp to 40% of total scoreNot included
QuestionsA mix of essay and multiple choice questionsA mix of essay, multiple choice, fill in the blanks, yes-no, true-false questions
Materials of examinationPresented materials and additional referencesPresented material
Final examinationUp to 50% of total scoreUp to 70% of total score
QuizNot included (0%)Not included (0%)

Which of the evaluation plans would you like to choose (please tick just one box) plan A ◻ plan B ◻

Finally, based on first phases of the study, the questionnaire of this study was developed. The questionnaire had two sections. First section included demographic, descriptive variables, and opinion of participants about attributes of evaluation methods. Second one included choice sets and definition of the attributes.

Data analysis

The base for analysis the DCE data is the random utility model (RUM). In according to RUM, in a situation that we request a person to choose between two evaluation plan of A or B, when he/she chose option A, it shows that this alternative have higher utility than alternative B. It can be write as follows:[28] U (A, C) > U (B, C)     (1) Here, U devotes the derived utility from the evaluation plans A and B. C is personal characteristics that are a common factor in two sides of equation 1, then the formula can be rewrite as: V (A-B) = U (A, C)– U (B, C)     (2) Here, V is indirect utility obtained from evaluation plan A in compare with evaluation plan B. the following linear regression model fitted to estimate the utility of attributes of evaluation plans: V = ß1dis+ ß2 mid+ ß3que+ ß4que+ ß5que+ ß6 sour+ ß7 final+ß8 quiz + ɛ     (3) Where, β1 to β8 are the coefficients of the attributes of evaluation plans: Dis (student's discipline), mid (midterm examination), que (question types), final (final examination), and quiz (quiz). ε devotes error term. With considering the logistic distribution for error term, the conditional logit regression model was performed for data analysis. The restrictive assumption of this model is independent of irrelevant alternatives. However, the interpretations of coefficients of this model is easier than other complicated models, such as nested or mixed logit models and it is shown this method is enough strong to use for analysis DCE data.[29] We performed a subgroup analysis to elicit preferences of students by schools. Goodness of fit of models was investigated using McFadden R2 and χ2 tests. The statistical software packages SAS 9.1 and Stata 14 (StataCorp, College Station, TX, USA) were used to perform the analysis.

Sampling

The sample size of the DCE studies is complex to determine. At least 30 participants should include in the studies. However, others suggested some formula to determining the sample size as follow:[30] Where, S is sample size, n is the total number of participants, t is the total number of choice sets, a is the number of alternatives in the choice sets, and c is the highest number levels in the included attributes. Accordingly, we included 213 students and 30 professors in the study. The convenience sampling method used for data collection by interview with participants.

Ethical statements

This study has been approved by the Ethics Committee of KUMS (ethics code: IR.KUMS.REC.1396.633).

Results

The descriptive results of the study showed that mean (standard deviation: [SD]) of students was 22.55 (2.49) years. Of total students, 213 (68.27%) were female and others were male. Samples were included from all schools of health (28.53%), nursing and midwifery (17.95%), medicine (4.49%), paramedical sciences (21.79%), pharmacy (6.41%), nutrition sciences (4.17%), and dentistry (16.67%). The mean (SD) age of professors was 37.93 (7.18) years and 76.67% were male. In addition, mean (SD) of number of courses that professors teach during their employment in the university was 108 (143.70). The highest percentage of professors (70%) had philosophy of doctor (PhD) degree; follow by degrees of master of sciences (26.67%), and medical doctor (3.33%). Figure 1 shows the percentage of agreement with including different types of questions in evaluation of students by professors and students. Professors believed that essay (80%), multiple choice (80%), and matching (76.67) questions were the most popular types of questions. On the other hand, multiple choice (84.62%) was the most preferred question by students and other types of questions were not very popular.
Figure 1

Percentage of agreement with inclusion of different types of questions in examinations

Percentage of agreement with inclusion of different types of questions in examinations Preferences of participants about related items to the evaluation of students showed that 13.33% of professors and 33.65% of students were not agree to allocate some scores to student's discipline in class. In case of midterm examination, the most of professors (36.67%) prefer to allocate up to 30% of evolution scores to this item. However, the most percentage of students (30.45%) were agree to include midterm examination up to 15% of total scores. The majority of students prefer to examination questions compromise just presented materials, while 70% of professors prefer to include additional text for evaluation examinations in theoretical courses. Table 1 shows preferences of participants about attributes of evaluation plans in theoretical courses.
Table 1

Preferences of participants about attributes related to evaluation plans of students

Evaluation itemsProfessors, n (%)Students, n (%)
Student’s discipline
 No4 (13.33)105 (33.65)
 Up to 5% of total score18 (60)124 (39.74)
 Up to 10% of total score8 (26.67)83 (26.6)
Midterm exam
 No2 (6.67)68 (21.79)
 Up to 15% of total score9 (30)95 (30.45)
 Up to 30% of total score11 (36.67)79 (25.32)
 Up to 40% of total score8 (26.67)70 (22.44)
Materials of exam
 Presented sources in class9 (30)269 (86.5)
 Presented sources in class + additional references21 (70)42 (13.5)
Final exam
 Up to 50% of total score13 (43.33)108 (34.62)
 Up to 70% of total score12 (40)144 (46.15)
 Up to 90% of total score5 (16.67)60 (19.23)
Quiz
 No4 (13.33)154 (49.36)
 Up to 5% of total score21 (70)120 (38.46)
 Up to 10% of total score5 (16.67)38 (12.18)
 Total30 (100)312 (100)
Preferences of participants about attributes related to evaluation plans of students The results of DCE model to eliciting preferences of professors showed that in univariate logit regression model midterm examination, type of questions (at the level of a mix of essay, multiple choice, fill in the blanks, yes-no, true-false), source of examination, and quiz significantly increased the probability of selecting evaluation packages. However, multivariate model for professors' preferences indicated that all attributes increased the probability of choosing an evaluation method. With increasing a percent of midterm scores, the probability of selecting an evaluation method increased by 2.2% (P < 0.01). The attributes of question types increased the utility of professors so that mix types of questions increased the probability of choosing the evaluation method by 17.7% than essay (P > 0.1). In case of source of examination, presented sources and additional references increased the utility of professors than presented sources in class (OR: 2.424, P < 0.01). Quiz, also, increased the utility of an evaluation package in professors. One percent increase in quiz share in total score increased the probability of selection an evaluation method by 7.3% (P < 0.01). The fitted model was statistically significant (P < 0.001). [Table 2] presents the results of DCE analysis for eliciting the preferences of professors for attributes of evaluation plans in theoretical courses.
Table 2

Preferences of professors about attributes of evaluation plans using discrete choice experiment analysis

AttributesMultivariate analysisUnivariate analysis


aOR95% CIPOR95% CIP
Student’s discipline1.0230.945-1.1080.5681.0270.979-1.0870.265
Midterm examination1.0221.007-1.037<0.0101.0311.018-1.048<0.001
Questions (ref: essay)
 Multiple choice0.7010294-1.6670.4210.5140.298-0.8880.017
 Essay and multiple choice1.1770.654-2.1160.5860.7390.447-1.2230.240
 A mix of essay, multiple choice, fill in the blanks, yes-no, true-false1.1770.493-2.8090.7141.1260.658-1.9310.666
Materials of exam (ref: presented material)
 Presented material and additional references2.4241.396-4.211<0.0102.3071.592-3.346<0.001
Final examination1.0040.984-1.0240.6890.9890.979-0.9990.043
Quiz1.0731.021-1.128<0.0101.0641.019-1.111<0.010
Observation540-
LR χ2 (8)51.27-
P<0.001-
McFadden R20.104-

aOR=Adjusted odds ratio, OR=Odds ratio, CI=Confidence interval, LR=Likelihood ratio

Preferences of professors about attributes of evaluation plans using discrete choice experiment analysis aOR=Adjusted odds ratio, OR=Odds ratio, CI=Confidence interval, LR=Likelihood ratio Univariate analysis of attributes of evaluation method showed that student's discipline and adding additional references to presented material in class have disutility for students (P < 0.001). Multivariate regression analysis indicated that student's discipline (P < 0.001), type of questions (P < 0.05), and quiz (P < 0.001) decrease the probability of selecting the evaluation method by the students. A mix of essay, multiple choice, fill in the blanks, yes-no, true-false questions in an evaluation method decreased the probability of choosing that method by 76%. One percent increase in final examination's score significantly increased the probability of the selecting evaluation method by 2.5% (P < 0.01). The fitted model was statistically significant (P < 0.001). Table 3 presents the results of DCE analysis for eliciting the preferences of students for attributes of evaluation plans in theoretical courses.
Table 3

Preferences of students about attributes of evaluation plans using discrete choice experiment analysis

AttributesMultivariate analysisUnivariate analysis


aOR95% CIPOR95% CIP
Student’s discipline0.9130.875-0.953<0.0010.9730.960-0.986<0.001
Midterm examination1.0091.003-1.0160.0071.0000.997-1.0040.559
Questions (ref: essay)
 Multiple choice0.5870.348-0.9880.0451.6711.436-1.945<0.001
 Essay and multiple choice0.5680.329-0.9800.0421.5171.311-1.776<0.001
 A mix of essay, multiple choice, fill in the blanks, yes-no, true-false0.2400.106-0.545<0.0011.0140.867-1.1870.857
Materials of examination ( ref: presented material)
 Presented material and additional references0.9840.853-1.1360.8280.8030.738-0.874<0.001
Final examination1.0251.010-1.0400.0011.0020.999-1.0050.114
Quiz0.9470.919-0.975<0.0011.0060.993-1.0200.357
Observation4368-
LR χ2 (8)93.930-
P<0.001-
McFadden R20.031-

aOR=Adjusted odds ratio, OR=Odds ratio, CI=Confidence interval, LR=Likelihood ratio

Preferences of students about attributes of evaluation plans using discrete choice experiment analysis aOR=Adjusted odds ratio, OR=Odds ratio, CI=Confidence interval, LR=Likelihood ratio Eliciting of the preferences of students by schools of the university showed that students not prefer to discipline have some scores for all schools. A mix of essay, multiple choice, fill in the blanks, yes-no, true-false questions had disutility for students of all schools. Regarding the source of examination, the level of presented material and additional references made disutility for students of nursing and midwifery, and dentistry schools. Finally, quiz had disutility for students from all schools. The results of DEC analysis for eliciting the preferences of students for attributes of evaluation plans in theoretical courses by schools are presented in Table 4.
Table 4

Preferences of students about attributes of evaluation plans using Discrete choice experiment analysis by schools

AttributesHealthNursing and midwiferyMedicineParamedical sciencesPharmacyNutrition sciencesDentistry







OR95% CIOR95% CIOR95% CIOR95% CIOR95% CIOR95% CIOR95% CI
Student’s discipline0.8750.791-0.9670.9300.848-1.0190.9280.765-1.1250.9730.880-1.0760.8460.682-1.0500.9950.797-1.2400.8810.800-0.969
Midterm examination1.0060.990-1.0211.0100.996-1.0241.0300.998-1.0621.0030.987-1.0181.0361.002-1.0700.9930.959-1.0271.0110.996-1.025
Questions (ref: essay)
 Multiple choice0.3600.102-1.2700.6520.216-1.9621.0000.098-10.1661.0000.287-3.4791.2000.087-16.4401.3330.088-20.1080.3550.112-1.123
 Essay and multiple choice0.3070.079-1.1760.8830.287-2.7211.0010.090-11.0821.0380.277-3.8860.5840.036-9.5071.7360.101-29.9440.3400.107-1.077
 A mix of essay, multiple choice, fill in the blanks, yes-no, true-false0.1050.015-0.7310.4770.083-2.7550.6550.016-26.1340.3880.056-2.6910.4440.007-25.7480.4950.007-34.3010.1320.021-0.818
 Materials of examination (ref: presented material)
 Presented material and additional references1.1180.852-1.4670.6900.486-0.9771.2130.560-2.4521.0250.753-1.3941.0900.545-2.1801.0670.522-2.1770.9600.666-1.383
Final examination1.0441.007-1.0801.0140.983-1.0461.0060.941-1.0741.0160.981-1.0531.0050.933-1.0811.0090.935-1.0891.0310.998-1.065
Quiz0.9320.868-1.0000.9500.893-1.0110.9470.827-1.0830.9670.901-1.0380.8650.747-1.0030.9950.852-1.1620.9540.896-1.015
Observation1246784196952280182728
Log likelihood−414.60−258−62.12−316.17−78.16−58.64−242.25
LR χ2 (8)34.47027.330.16927.5437.758.8720.1
P<0.001<0.0010.169<0.001<0.0010.354<0.01
McFadden R20.0400.050.0850.0420.1940.0700.039

aOR=Adjusted odds ratio, OR=Odds ratio, CI=Confidence interval, LR=Likelihood ratio

Preferences of students about attributes of evaluation plans using Discrete choice experiment analysis by schools aOR=Adjusted odds ratio, OR=Odds ratio, CI=Confidence interval, LR=Likelihood ratio

Discussion

This study primarily revealed that professors think essay, multiple choice, and matching questions are more appropriate than other questions for the evaluation of students' knowledge. They less prefer using other questions such as the fill in the blanks, yes-no, and true-false questions. On the other hand, students prefer multiple questions much more than other question types. Thus, it could be concluded that MCQ could be an evaluation tool that prefer by both students and professors. MCQs are using widely in universities and entrance examinations.[3132] They have many pros and cons; but, if they are not designed well, barley considered as a suitable tools for deep knowledge assessment.[33] On the other hand, some studies revealed are a suitable assessment for in-depth learning, if they are designed properly.[3234] Another issue about MCQs is the chance for cheating. Since this study revealed that MCQs are much more popular for students in compare with other types of questions, this maybe because of their easiness for cheating.[35] Now, there are several ways that could reduce the probability of cheating in MCQs such as test randomization,[36] computer-aided assessment,[37] and automatic test generator.[38] The results of a study for investigating evaluation methods in theoretical and practical courses of laboratory sciences students in Tehran University of Medical Sciences showed that the midterm exam is used in 35% of the total theoretical and practical courses, and among the evaluation methods, the MCQs method with 70% was among the common methods of assessments.[22] The study of Kojury et al. (2017), about types of evaluation methods of medical students at the Shiraz University of Medical Sciences indicated that MCQs is the most frequent method of evaluation. Based on this fact that MCQs cannot evaluate higher level of learning in students, this study suggested a mix of different types of evaluation methods such as descriptive examinations, and Mini-CEX in order to evaluate higher level of cognitive knowledge.[39] In brief, in medical sciences assessment should be done on knowledge and skills of the student. However, MCQs fail to assess students' skill.[40] Thus, the limitations of MCQs should be considered by universities and professors who use it. It could be concluded that all types of question especially essay should be mixed with MCQs for better evaluation. Furthermore, evidence suggests that viva voce or oral examination method is a valid and reliable technique of student evaluation. This method can deeply evaluate the knowledge and skills of students and provide more opportunity for dialectic communication between examiner and student. However, examiner should manage stress of students for increasing effectiveness of this method.[41] Many variables can effect on students preferences about attributes of evaluation methods. According to the current knowledge, personality and gender are among the main factors that influence individuals' preferences for student evaluation methods, including the type of questions.[42] In addition, gender, age, and emotional variables are known as variables that effect on academic achievement. Females more probability to have higher commitment and higher final degree grade point average compared to male students. In comparison to extraverts' personality, openness students favor essays and oral exams and do not prefer multiple choice and group work. In addition, females prefer coursework evaluation and oral examinations are more favorable for males. Besides personality, learning style, and intelligence supposed to be associated with preferences of student for examination techniques.[43] According to our findings, more students preferred to discipline in class not include in the evaluation process, while more than 75% of professors agreed to consider the discipline of students as an item in this process. In addition, students had a higher tendency to have a midterm exam. About half of the students did not prefer to have a quiz in their evaluation. According to DCE analysis, professors preferred to student's discipline, final examination, and Quiz have higher scores in evaluation of students. Furthermore, they rather a mix of essay, multiple choice, fill in the blanks, yes-no, true-false questions to include in the evaluation exams. Moreover, they tended to consider presented material in the courses and additional related references as sources of examinations. On the other hand, considering a proportion of total score to discipline of students, additional sources for examinations, and quiz by professors for evaluation made disutility for students. Furthermore, students prefer essay and MCQs than other type of questions. MCQs made higher utility for pharmacy, and nutrition sciences schools' students than other ones. Similarly, another study in Iran showed there are some match and mismatch points between learning preferences of professors and students.[44] Aragon (2010) indicated that if face-to-face and online training courses are designed appropriately and based on learning theories, they will equally affect people's learning. Therefore, carefully designing training courses and evaluating tests and adapting these courses to students' preferences can play an important role in learning and evaluating them.[45] One of the strength of this study is employing an econometric technique to find students and professors which deal with the research novelty. It is the first attempt in Iran to applying this method in eliciting preferences in medical education. Another strength is considering students' and professors' preferences simultaneously. The research achieved sound knowledge about educational assessment in universities. However, it faced with several limitations. The first limitation is related to sample selection method which was convenient sampling. Thus, the results may not be completely generalizable to study population or other universities. Study participants were included from KUMS. Therefore, the results could not representative of preferences of professors and students in other medical university of Iran. As there are limitations for number of included attributes in DCE method, we could not consider all important attributes. Therefore, future works should consider other aspects of evaluation methods.

Conclusion

The findings showed MCQs as a favorite test form for both professors and students. However, professors prefer other types of question including essays and matching questions too. Both professors and students prefer a mix of questions in the exams. There is a gap between preferences of professors and students regarding attributes of evaluation methods such as student's discipline, examination materials, and quiz. Further studies are needed to examining other attributes of evaluation methods in theatrical and practical courses in Iran.

Financial support and sponsorship

This study funded by Deputy of research and technology of KUMS (fund No. 96641).

Conflicts of interest

There are no conflicts of interest.
  13 in total

1.  A progressive nurse practitioner student evaluation tool.

Authors:  Tamera Pearson; Linda Garrett; Susan Hossler; Peggy McConnell; Jennie Walls
Journal:  J Am Acad Nurse Pract       Date:  2012-04-30

2.  Practical strategies for nursing education program evaluation.

Authors:  Lynne Porter Lewallen
Journal:  J Prof Nurs       Date:  2014-09-10       Impact factor: 2.104

3.  Are Multiple Choice Questions for Post Graduate Dental Entrance Examinations Spot On?-Item Analysis of MCQs in Prosthodontics in India.

Authors:  Kamal Shigli; Sushma S Nayak; Sivaranjani Gali; Banashree Sankeshwari; Deepti Fulari; K Shyam Kishore; Nagaraja Upadhya P; Vasanti Jirge
Journal:  J Natl Med Assoc       Date:  2017-12-06       Impact factor: 1.798

4.  Manganese (II) and substrate interaction with unadenylylated glutamine synthetase (Escherichia coli w). II. Electron paramagnetic resonance and nuclear magnetic resonance studies of enzyme-bound manganese(II) with substrates and a potential transition-state analogue, methionine sulfoximine.

Authors:  J J Villafranca; D E Ash; F C Wedler
Journal:  Biochemistry       Date:  1976-02-10       Impact factor: 3.162

5.  Statistical Methods for the Analysis of Discrete Choice Experiments: A Report of the ISPOR Conjoint Analysis Good Research Practices Task Force.

Authors:  A Brett Hauber; Juan Marcos González; Catharina G M Groothuis-Oudshoorn; Thomas Prior; Deborah A Marshall; Charles Cunningham; Maarten J IJzerman; John F P Bridges
Journal:  Value Health       Date:  2016-05-12       Impact factor: 5.725

6.  Modeling the problem-based learning preferences of McMaster University undergraduate medical students using a discrete choice conjoint experiment.

Authors:  Charles E Cunningham; Ken Deal; Alan Neville; Heather Rimas; Lynne Lohfeld
Journal:  Adv Health Sci Educ Theory Pract       Date:  2006-08       Impact factor: 3.853

7.  Constructing experimental designs for discrete-choice experiments: report of the ISPOR Conjoint Analysis Experimental Design Good Research Practices Task Force.

Authors:  F Reed Johnson; Emily Lancsar; Deborah Marshall; Vikram Kilambi; Axel Mühlbacher; Dean A Regier; Brian W Bresnahan; Barbara Kanninen; John F P Bridges
Journal:  Value Health       Date:  2013 Jan-Feb       Impact factor: 5.725

8.  Usefulness of 360 degree evaluation in evaluating nursing students in Iran.

Authors:  Tabandeh Sadeghi; Marzeyeh Loripoor
Journal:  Korean J Med Educ       Date:  2016-02-25

9.  Developing attributes and levels for a discrete choice experiment on basic health insurance in Iran.

Authors:  Ali Kazemi Karyani; Arash Rashidian; Ali Akbari Sari; Sara Emamgholipour Sefiddashti
Journal:  Med J Islam Repub Iran       Date:  2018-03-28

10.  Evaluating the clinical quality of departments as viewed by juniors and seniors of Shiraz dental school.

Authors:  Shahla Momeni Danaei; Elham Mazareie; Sahar Hosseininezhad; Mahsa Nili
Journal:  J Educ Health Promot       Date:  2015-12-30
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.