The literature is mixed on whether evaluation and feedback to clinical teachers improves clinical teaching. This study sought to determine whether resident-provided numerical evaluation and written feedback to clinical teachers improved clinical teaching scores.
Anesthesia residents anonymously provided numerical scores and narrative comments to faculty members who provided clinical teaching. Residents returned 19,306 evaluations between December 2000 and May 2006. Faculty members received a quantitative summary report and all narrative comments every 6 months. Residents also filled out annual residency program evaluations in which they listed the best and worst teachers in the department.
The average teaching score for the entire faculty rose over time and reached a plateau with a time constant of approximately 1 yr. At first, individual faculty members had average teaching scores that were numerically diverse. Over time, the average scores became more homogeneous. Faculty members ranked highest by teaching scores were also most frequently named as the best teachers. Faculty members ranked lowest by teaching scores were most frequently named as the worst teachers. Analysis of ranks, differential improvement in scores, and a decrease in score diversity effectively ruled out simple score inflation as the cause for increased scores. An increase in teaching scores was most likely due to improved teaching.
A combination of evaluation and feedback, including comments on areas for improvement, was related to a substantial improvement in teaching scores. Clinical teachers are able to improve by using feedback from residents.
What We Already Know about This Topic
❖ Feedback is important to improved teaching, yet there are no long-term studies examining resident feedback on anesthesiologist teaching.
What This Article Tells Us That Is New
❖ Over a 5-yr period, residents provided qualitative and quantitative anonymous evaluations of teaching faculty.
❖ Institution of this feedback system was associated with increased teaching scores for the faculty.
RESIDENCY programs aspire to improve clinical teaching provided by clinician educators. One strategy used to improve clinical teaching is to obtain resident evaluations of the teachers. To date, evaluations provided by residents and medical students remain most common. The effect of evaluations on teaching has been mixed. Some studies demonstrate an increase in clinical teaching scores after written feedback,1–3whereas feedback in the form of simple numerical ratings results has not improved teaching scores.4–8Some studies have been underpowered to find a difference in teaching scores,7,9,10whereas others may fail to show improvement because of a ceiling effect.7,9,11The literature is also largely silent on the issue of the time needed to improve clinical teaching. Concerns have been raised about the reliability and validity of resident and student evaluations of both clinical and classroom teaching.12,13Additional evidence is needed demonstrating that resident evaluation and feedback either does or does not lead to durable improvements in clinical teaching.
Feedback is fundamental to performance improvement.14Recent studies have repeatedly shown that self-assessment can be remarkably flawed, with the worst performers most seriously overestimating their skills.15–18Claridge et al .19studied surgeon self-evaluation of teaching and compared it with resident evaluation of their teaching. None of the surgeons who received below average teaching scores self-identified these deficiencies. It is noteworthy that faculty members who declined to engage in self-evaluation had lower teaching scores than those who volunteered to participate. The positive effects of immediate feedback to lecturers was demonstrated by improved teaching scores after second-year medical students provided feedback with numerical ratings and narrative comments.20
The current study provides a long-term (5.5 yr) examination of the influence of resident evaluation and feedback on the clinical teaching faculty and strongly supports the conclusion that resident evaluation and feedback increases clinical teaching scores. The report includes data on the kinetics of improvement and a novel form of construct validity relating to teaching scores. The data also effectively rule out the possibility that teaching scores increased as a result of simple grade inflation of the scores given by our residents.
Materials and Methods
Resident Evaluation and Feedback Regarding Clinical Teaching.
We developed an evaluation system to capture anonymous resident feedback regarding faculty members engaged in intraoperative and perioperative clinical teaching. Each month, our computerized billing database determined which resident has worked with which attending physician. For rotations without billing information (Obstetrics, Intensive Care Unit, Pain Rotation, Preadmission Testing Area, and Postanesthesia Care Unit), we used schedules to create the resident-attending physician matches. Each unique resident-attending physician pair results in a request for the resident to anonymously evaluate that faculty member. The paper-based evaluation form lists seven different attributes of teaching: overall, time spent, clinical supervision, quality of teaching, quantity of teaching, role model, and encourages thinking about the science of anesthesia. Each question was rated using a Likert scale ranging from 0 to 10. Zero denotes the worst teaching and 10 denotes the best teaching. Teaching scores are formed by summing up the seven subscores, and thus teaching scores ranged from 0 to 70. Each form requests narrative comments in three areas—strengths, areas that need improvement, and additional comments. Residents were told that whenever they give low scores that they should provide a specific comment regarding what they would like the faculty member to start doing, do more, or stop doing to improve their teaching. During the last 2 months of each 6-month period, residents who had not completed any evaluations were contacted by letter and encouraged to complete and submit their evaluations. Approximately 89% of our clinical rotations occur at the Massachusetts General Hospital. Evaluations pertain only to Massachusetts General Hospital faculty members. The Massachusetts General Hospital Institutional Review Board waived the need for informed consent and classified this study as exempt.
Analysis and Reports of Faculty Member Teaching.
Numerical results and verbatim comments from each evaluation were keyed into an electronic database by a single person. Every 6 months, the data were analyzed, and individual reports were prepared for each faculty member who had at least two evaluations. The report provides the faculty member with an average score for each of the seven areas of teaching. They are also provided with an overall composite teaching score, which is the sum of the seven subscores. Reports contain the teaching score of the “average faculty member.” The “average faculty member” is represented by the average of all data collected during the 6-month period and includes the average score for each of the seven subscores as well as the overall composite teaching score. Any significant differences between the individual and the average faculty member were highlighted for both subscores and the overall composite teaching score. Reports also contained a graphical comparison of the individual faculty member's composite score compared with all other faculty members. Resident comments pertaining to individual faculty members were included with each individual report.
Each 6-month time window is referred to as a period. Each period was numbered sequentially. Period 1 was our initial or baseline use of this evaluation system and refers to the 6-month time window from December 1, 2000 to May 31, 2001. Period 2 refers to the subsequent 6 months and so forth. During the first six periods, individual reports also contained the explicit rank of each attending physician (e.g. , rank 33 of 125). Explicit reporting of rank was stopped after period 6 because scores were so similar that rank differences were largely meaningless. Relative ranks for a period were determined by dividing a rank number by the total number of faculty evaluated in that period. Thus, relative ranks begin near 0 (top ranked person) and progress to 1 (lowest ranked person). Relative ranks allow rank positions to be compared across periods having different numbers of faculty. During the first six periods, faculty members were asked to speak with the chairman if they had both very low scores and negative comments. These few faculty members were encouraged to improve their teaching by meeting with a single senior faculty member who was experienced in faculty development and education. Approximately three to four faculty members per period took advantage of this offer, but we have no formal record of this activity because their involvement was voluntary. Teaching scores and comments were not otherwise used for individual faculty development programs, entitlements (i.e. , travel or academic time), or teaching assignments. We did use teaching scores in part to help identify the best teachers to guide annual bonus distribution. The faculty was not made aware of the metrics that went into this decision, and this was not a formalized program. Except for the reports that were distributed every 6 months, faculty members received no further information regarding their teaching. Thus faculty members were provided repeated rounds of feedback and essentially were allowed to decide for themselves how to interpret the results and how to improve their teaching.
End-of-Year Resident Survey Listing Best and Worst Teachers.
Toward the end of each academic year, we anonymously survey our residents regarding a wide variety of issues. Among the questions is a request to list the best and worst teachers in the department. The number of times a faculty member was listed was converted into a frequency histogram and plotted as a function of that person's relative rank as determined using teaching scores over that same academic year. Histogram counts were determined independently from teaching scores.
Scores in different periods are compared by way of unpaired t tests assuming unequal variances. Exponential fits were determined using a Levenberg-Marquardt method. All statistics were determined using StatsDirect (version 2.6.6; StatsDirect Ltd., Cheshire, United Kingdom), Excel 2003 (Microsoft, Redmond, WA), or Origin (version 7.5 SR4; OriginLab Corp., Northampton, MA). Effect sizes were determined by Cohen d values which are calculated as the difference in means divided by the combined SD of the data. Effect sizes provide a measure of the size of a difference compared with the variation in the data. Effect sizes are classified as small (Cohen d = 0.2), medium (Cohen d = 0.5), or large (Cohen d = 0.8).21Cronbach α was used to examine reliability between subscores. Rank data are compared using Kendall τ. Kendall τ is used to determine whether two rank orders are the same. When two rank orders are identical, Kendall τ is 1.0; if they are perfectly inversely related, it is −1.0; and if the rank orders bear no relationship to each other, then it is 0.0. P values are two-sided and determined exactly whenever possible. Data points in graphs are means ± SEM unless noted otherwise.
Teaching Scores Increased after Implementing an Evaluation and Feedback Process
During the 5.5 yr of this study, a total of 19,306 evaluations were returned by 194 different residents concerning 197 different faculty members. Table 1shows the number of evaluations, residents, and faculty members during each 6-month period. The overall Cronbach α measure for internal consistency for all seven subscores over 19,306 evaluations was 0.980. This high Cronbach α strongly suggests that residents generally use each of the subscales in an interchangeable way; thus, it is unlikely that there is enough unique information in the individual subscores to allow meaningful comparisons. Because the subscales were used only to compute a single teaching score, they were not analyzed further.
All individual teaching scores are shown for periods 1 and 7 (fig. 1). Period 1 represents the baseline distribution of scores, and period 7 is representative of all later periods. Teaching scores during period 1 decreased over the first 80% of the faculty (until relative rank 0.8) and then declined more rapidly. In period 7, approximately 3 yr later, teaching scores declined less rapidly as one went down the rank order. The mean teaching score in period 7 was higher than in period 1. The average teaching score increased from period 1 up until approximately period 6 (fig. 2A, solid circles). The overall difference in teaching scores between periods 1 and 6 was significant (P = 5 × 10−37). The effect size, Cohen d , for the change in scores between periods 1 and 6 was 0.50, a medium-sized effect. On a 0–10 scale, this corresponds to a change of 0.8 (from 7.8 up to 8.6). To remove concern that the increase in teaching scores was due to a changing faculty composition, only the scores of the 50 faculty members who were present for all 11 periods were examined. The teaching scores of these 50 persistent faculty members were very similar to the overall teaching scores for all faculty members (fig. 2A, +).
The Time Course for Improvement
The time frame over which the teaching scores increased was well described by an exponential curve with a time constant of 0.94 ± 0.11 yr (fig. 2A). The change in teaching scores was nearly complete after three time constants, which corresponds to period 6 or 7.
Faculty Members' Scores Became More Homogenous
Over the same time frame that the overall teaching scores were increasing, faculty members' average scores were becoming more similar. In period 1, the teaching scores were broadly distributed and covered a wide range as the relative rank order is descended (fig. 1). In contrast, in period 7, individual average scores occurred over a much narrower range as the relative rank order is descended (fig. 1). The spread in scores was quantified by determining the SD of the average teaching scores of the faculty members in each period. Figure 2Bshows the SD of the faculty member's scores as a function of period. The diversity of scores decreases exponentially with a time constant of 1.21 ± 0.37 yr. The change in score diversity (as represented by the group SD) is largely complete by approximately three time constants, which approximately corresponds to period 7. The difference in group score diversity between periods 1 and 9 was significant as determined by an F test on the group variance (P = 4.9 × 10−11). The effect size, Cohen d , for the change between periods 1 and 9 was 0.95, a large effect.
An Independent Determination of Teaching Quality
Our yearly anonymous residency program evaluations ask residents to list (with no limit) the best and worst clinical teachers. This provides an independent and nonnumeric approach to assessing teaching quality. We do not provide this information to the faculty and thus it has no impact on them. The number of times that a faculty member is listed as a “best” or “worst” clinical teacher was counted for each academic year, and these frequency data were plotted against the corresponding relative ranks for these same faculty members over these same time periods. Figure 3shows that faculty members who had the highest counts for best teacher were also independently ranked the highest using numerical teaching scores. Likewise, faculty members who had the most counts for worst teacher were also ranked lowest based on our numerical teaching scores. The “best” histogram was fit by an exponential function that decayed with a “relative rank” rate of 15.3 ± 0.02%. This means that for every 15.3% reduction in relative rank, the number of times a faculty member was labeled as “best” was reduced by 63%. Thus, after three relative rank rates (which encompassed the top 46% of the faculty), it became unlikely that a faculty member was labeled as one of the “best” teachers. The “worst” histogram was fit by an exponential function that decayed with a relative rank rate of 16.6 ± 0.03%. This means that for every 16.6% increase in relative rank, the number of times a faculty member was labeled as “worst” was increased by 63%. Thus, it became more likely that a faculty member was labeled as one of the “worst” teachers as their relative rank increased and especially increased as they fell into the bottom half of the relative ranks. It is noteworthy that even faculty members who were ranked in the lower half of the numerical relative ranks were sometimes listed among our best faculty. The faculty listed as the worst teachers mainly dwell within the lowest 20% of the numerical relative ranks. Overall, the residents listed 870 names as “best” and 132 names as “worst.” Thus, the number of teachers listed as “best” was more than 6 times greater than the number listed as “worst.”
Did Residents Systematically and Indiscriminately Increase Teaching Scores?
If residents systematically and indiscriminately provided higher teaching scores for any reason (Grade Inflation Model), then as teaching scores increased, the rank order of the faculty would remain the same; scores would increase equally for all faculty members, diversity of scores would remain the same, and baseline scores given by residents would increase.
Rank Order Was Not Preserved over Time
To examine the stability of rank orders over time, the 50 faculty members who were evaluated in all 11 periods were studied. Their teaching scores are representative of the entire faculty (fig. 2A, +). They were ranked from 1 to 50 for each of the periods 1–11 based on their teaching scores during each period. When the rank order from period 1 was compared with any other later rank order (periods 2–11), the average Kendall τ was 0.42. When comparing the rank order of period 1 to any other later rank order, the maximum Kendall τ was only 0.61 (period 1 vs. period 11), and the mean upper 95% confidence interval for the Kendall τ was 0.57. Kendall τ never reached 1, which implies that later ranks had significant differences from the baseline rank order of period 1. Thus, rank order significantly changed as teaching scores increased over time.
The teaching score distributions shown in figure 1reveal that teaching scores are not linearly distributed over the entire rank order. Teaching scores are disproportionately high and low in the top and bottom quarter of the rank list. Thus the faculty in the top and bottom quarter appeared separate from the middle faculty. Teaching scores from the middle 50% of the ranks is quite linear (period 1, rank region 0.25–0.75 r 2= 0.99, P < 0.0000001; period 7, rank region 0.25–0.75 r 2= 0.99, P < 0.0000001). We used this finding to divide the 50 persistently present faculty members into three categories: top 25%, middle 50%, and bottom 25%, which corresponds to the top 12, middle 26, and bottom 12 faculty members. When the top and bottom groups were analyzed for stability of rank order, they were notably better preserved than were the ranks of the middle 26 (fig. 4A). In particular, the faculty who initially occupied either the top 12 or the bottom 12 ranks in period 1 retained much of their rank order into periods 8, 9, 10, and 11. A significant relationship was found between the early rank order (period 1) and each of the later rank orders (period 8, 9, 10, or 11). The average Kendall τ for these rank-order comparisons was 0.51 (P < 1 × 10−7), and the 95% confidence interval did not include either 0 or 1 (fig. 4B). This contrasts with the complete mixing of rank orders for the middle 26 faculty members between period 1 and periods 8–11 (fig. 4A). The average Kendall's τ for these rank-order comparisons for the middle 26 faculty members was −0.064 and was not different from 0 (P = 0.34475) (fig. 4B). This analysis reveals that the middle 26 faculty members changed ranks to the extent that the initial rank order had no relationship to later rank orders.
Teaching Scores Increased the Most for Lower Ranked Faculty
The change in teaching scores between periods 1 and 9 for the top 12, middle 26, and bottom 12 ranked faculty members were computed. The scores increased the least for the top 12 ranked faculty members, moderately for the middle 26 ranked faculty members and the most for the bottom 12 ranked faculty members (fig. 5). The differences in score increases were all significant (top vs. middle, P = 0.0091; middle vs. bottom, P = 0.00016; top vs. bottom, P = 0.0000014). The increases in teaching scores are thus not equal for all faculty members and instead are rank-related.
Junior Residents' Scores Stayed the Same Whereas Senior Residents' Scores Increased
The teaching scores given by junior residents (1–4 months of residency) and senior residents (24–36 months of residency) are plotted as a function of period (fig. 6). The scores given by junior residents did not change over time (P = 0.99, fig. 6A). In contrast, the scores given by senior residents increased over time (P = 0.0021, fig. 6B).
Did Clinical Teaching Scores Improve?
The data from this study provide evidence that resident-based evaluation and feedback increases the clinical teaching scores of the teaching faculty. The increase in teaching scores was similar whether it was determined using all faculty evaluations in each period or just the scores from the 50 faculty members who were present throughout periods 1–11. This indicates that the increase in scores was not due to a change in faculty composition. Prior studies have found mixed results of the effectiveness of evaluation and feedback on teaching.1,3,7The current analysis differs from prior analyses by including a far larger number of evaluations. This study also provides the first measurement of the time needed to improve teaching. Few prior studies have had this longitudinal perspective or quantity of data.
A Novel Form of Construct Validity
We used an independent measure (being listed as “best” or “worst” teacher) to provide a non-numerical assessment of teaching. Our numerical teaching scores strongly indentify the same high and low performers as concurrently determined by counts of “best” and “worst” teacher designation. Overall, the histograms support our numerical evaluation system with concurrent construct validity.
Are Comments Necessary to Improve Teaching?
Although this study did not assess faculty members' interest in receiving feedback from residents, a prior study showed a strong interest of a volunteer community-based faculty in receiving feedback from medical students.22The community-based faculty valued student feedback over all other benefits offered to them for their teaching efforts, including money. This implies that at least some faculty members want feedback and may be interested in using it to improve their teaching. It is noteworthy that not all faculty members are interested in resident feedback about their teaching, and this can reach the level of resentment.10Our residents provided both quantitative (numeric) as well as qualitative (comments) feedback to the faculty. The comments provide direct constructive feedback for faculty members to use as tools to improve performance. Most faculty evaluation systems that fail to show improvement included only numerical ratings of the faculty without formative comments to help faculty improve.3–6In contrast, most faculty evaluation systems showing improvement in teaching scores, including the current study, included comments detailing strengths and areas for improvement.1–3,20The longest previous study of teaching scores did not show improvement over its 9-yr duration.5However, this study provided faculty only with numerical ratings and lacked formative comments. It is noteworthy that the addition of a comment section was associated with an improvement in scores within 1 yr.2Thus formative comments about strengths and weaknesses seem to be a key component enabling faculty to identify areas to improve. It is noteworthy that a study that provided only comments on areas of strength (with no mention of areas for improvement) showed no improvement over a 5-yr period.8Thus, our data suggest that areas of weakness need to be specifically identified to achieve increased teaching scores. Only when the teacher knows what areas to improve can they target the areas. Self-evaluation has proved to be remarkably inaccurate,16,17and thus external feedback provided by resident comments is likely to help identify areas that need improvement.
The impact of numerical-rating feedback versus comments-based feedback was determined using the studies with sufficient information (table 2). An overview of the studies in table 2reveals that eight studies, including the current one, were designed to look at teaching scores over time. Four studies showed improvement in faculty teaching scores and four studies showed no improvement. All four studies showing an improvement included comments in the feedback material. All four studies demonstrating no improvement included only numerical feedback. Among these eight studies, a chi-square test reveals that comments were related to improved scores (Fisher exact test, P = 0.03). Thus, comments seem to be the driver for improved teaching, whereas numerical ratings track teaching but do not drive improvement.
The Kinetics of Improved Clinical Teaching Scores
The time course for improving teaching scores was well described by an exponential function with a time constant of approximately 0.94 yr. Approximately 95% of the improvement occurred within three time constants, which corresponds to approximately 2.8 yr in the current study. The change in score diversity showed a very similar temporal change, with a time constant of 1.2 yr.
Our evaluation and feedback process is quite similar to the evaluation and feedback process used by Schum and Yindra1in their “feedback” group. Their faculty had received feedback every 2 months for a total of six episodes of feedback.1They found improvement in 4 of the 10 traits in the feedback group. The effect size of the findings of Schum and Yindra was 0.22 after 1 yr, although it was not statistically significant. Cohan et al. 2also had a process that was closely mimicked by the current evaluation and feedback process; they used a single annual feedback process and included specific suggestions from residents for ways that faculty could improve. His study found improvement in faculty teaching scores after 1 yr of feedback. Their data show an effect size of 0.31.2In the present study, at 1 yr, the faculty was provided with two episodes of feedback, and this resulted in an effect size of 0.26 (Cohen d of 0.26). In each case, after 1 yr, the magnitude of the improvement (the effect size) was similar. This suggests that the frequency of feedback may not be the limiting factor. Rather, it seems that faculty may take time to adjust their teaching skills and suggests that behavioral change is the actual rate-limiting step in improvement.
Do Clinical Teaching Scores Reflect Quality of Teaching?
Our teaching scores are believed to reflect actual quality of teaching. This hypothesis is based on the strong relationship between “best” and “worst” teachers and the concurrent teaching scores. When our residents list a faculty member as “best” or “worst,” it is highly likely that the numerical scores will be accompanied by a high or low clinical teaching score, respectively (fig. 3). We also allow our residents to choose a “Teacher of the Year,” and faculty members who are chosen regularly score in the top tier of our scoring system (data not shown). It is noteworthy that this form of evaluation synchrony is a form of “convergent validity”23and adds strength to our use of teaching scores to identify excellent teachers. Despite these relationships, we do not have any externally valid outcome data showing that the “best” teachers actually improve learning in the residents that they teach. This lack of outcome data is common in medical school and residency education.23Even where external raters have shown significant agreement with medical student ratings of lecturers,24there are not necessarily data showing that “better” lecturers result in “better” learning. Fortunately, there are examples of teaching scores being related to better student performance.25In university settings, student evaluations of teaching are associated with valid forms of achievement.26,27Although this study did not demonstrate improved learning with higher teaching scores, a body of literature demonstrates that better learning outcomes occur with higher teaching scores.25–29
How Do the Present Teaching Scores Compare to Others in the Literature?
Lectures and clinical teaching are usually evaluated using a Likert scale. Scores can be normalized by dividing the actual score by the dynamic range of the Likert scale. This converts each Likert score into a fraction of the maximum attainable score. Normalized teaching scores were computed from a variety of different teaching venues and found to be remarkably similar, with an overall mean of 0.797 (table 2). The 99% confidence interval for this mean was calculated as 0.774–0.820. During our baseline (period 1), our normalized clinical teaching score was 0.780, which falls within the estimated 99% confidence interval determined from the published studies. After our teaching scores had improved (period 7), our normalized clinical teaching score was 0.857, which is distinctly above and outside the 99% confidence interval for the mean estimated from the literature. Our evaluation and feedback system thus seems to have produced one of the highest normalized teaching scores reported.
Did Teaching Scores Increase As a Result of Simple Grade Inflation?
Simple grade inflation would increase teaching scores, leave the faculty rank order intact, increase scores equally for all faculty members, maintain score diversity, and increase the initial scores early in residency. Our data show that as teaching scores increased, the faculty rank order was not preserved (fig. 4), scores increased disproportionately for those whose ranks were lowest (fig. 5), scores became more homogeneous across faculty members (fig. 2B), and scores at the outset of residency were constant (fig. 6A). These analyses effectively rule out simple grade inflation as the cause for our increased clinical teaching scores.
The rank order changed over time, meaning that some but not all faulty members received higher scores, which in turn caused the rank order to change. The data also revealed that the top-ranked faculty members improved least, perhaps because of a ceiling effect. The middle- and lowest-ranked faculty members' scores improved more than the top-ranked faculty. In fact, the lower the rank, the more they improved (fig. 5). Although the lowest-ranked faculty members improved the most, their very low initial teaching scores caused them to remain ranked near the bottom. This manifested as increased scores but persistently low rankings. The finding that the lowest ranked faculty members improved the most has been reported before.1,2A reduced but persistent gap in performance between top and bottom performers has been reported in a prior study.2
Why Did Junior Residents Give Similar Teaching Scores over Time Whereas Senior Residents Gave Higher Scores over Time?
When residents first start in residency, they typically find every interaction with a faculty member educational. Junior residents are typically very pleased with the teaching they receive at the beginning of residency. This may explain why scores of junior residents remain stable and high over time. The consistent teaching scores given by beginning residents argues strongly against simple grade inflation over time.
As residents become more senior, they become better at discriminating various aspects of teaching,6which implies that they become more sophisticated “consumers” of clinical teaching. Tortolani et al .6showed that senior residents used teaching evaluations in a more complex fashion that their more junior counterparts.
In the early periods, the lower scores given by senior residents indicate that they had became progressively less satisfied with the teaching they received as they progressed through residency. As evaluation and feedback affected the faculty and teaching improved, the senior residents became more pleased with the clinical teaching that they received, and the decline of the scores disappeared.
Limitations of This Study
This study's primary limitations are lack of a control group and lack of outcome data showing that better teaching scores translate into better learning outcomes. The lack of a control group means that other variables, including the Hawthorne effect, may have led to improved teaching scores. However, the Hawthorne effect would probably cause simple grade inflation, and our data have ruled that out. Our baseline ranks in period 1 pose another limitation, in that some findings in this study relate to rank orders and whether they change. This presumes a stable baseline, something that was not demonstrated. However, our residency had not undergone any large changes during the few years leading up to period 1. Our initial data (period 1) was acquired, analyzed, and reported back to the faculty after all the evaluations for period 1 had been received. Thus they had no feedback until after the conclusion of period 1. There is no reason to expect that the faculty were acting on data that they had not yet received. Our findings are also limited to the context in which our residents receive clinical education. Our clinical teaching interactions involve a great deal of direct observation and supervision. This allows our residents an excellent opportunity to evaluate the faculty member's clinical teaching. We did ask a small number of the lowest ranked faculty to work with a senior faculty member with an interest in education. We have no formal measurement of the impact of this voluntary intervention because we do not know which faculty members chose to use this resource. The small number of faculty identified and the likely smaller number choosing to use this resource make it unlikely to have strongly influenced the results. Last, the cause for the lack of agreement in the middle parts of the rank order in period 1 in comparison with the rank order in later periods (8–11) could be due to a lack of precision in measuring the teaching scores for each individual. In particular, the ability to create a reliable rank ordering is reduced in the middle ranks, where teaching scores are similar.
The author thanks Eleanor Cotter, A.S. (Education Coordinator, Department of Anesthesia, Critical Care and Pain Medicine, Massachusetts General Hospital, Boston, Massachusetts), for accurately and confidentially transcribing every aspect of evaluation and feedback into electronic form.