SciELO - Scientific Electronic Library Online

 
vol.45 número2Cinema brasileiro e o ensino dos transtornos da personalidadeO uso de metodologias ativas de ensino-aprendizagem em pediatria: uma revisão narrativa índice de autoresíndice de assuntospesquisa de artigos
Home Pagelista alfabética de periódicos  

Serviços Personalizados

Journal

Artigo

Compartilhar


Revista Brasileira de Educação Médica

versão impressa ISSN 0100-5502versão On-line ISSN 1981-5271

Rev. Bras. Educ. Med. vol.45 no.2 Rio de Janeiro  2021  Epub 11-Jun-2021

https://doi.org/10.1590/1981-5271v45.2-20190340.ing 

Artigo Original

Clinical skills assessment and feedback in pediatric residency

Bruna de Sá Duarte Auto1 
http://orcid.org/0000-0001-7316-2914

Maria Viviane Lisboa de Vasconcelos1 
http://orcid.org/0000-0002-7861-7225

Ana Lydia Vasco de Albuquerque Peixoto2 
http://orcid.org/0000-0002-3893-0591

1 Universidade Federal de Alagoas, Maceió, Alagoas, Brazil.

2 Universidade Estadual de Ciências da Saúde de Alagoas, Maceió, Alagoas, Brazil.


Abstract:

Introduction:

A medical residency is the specialization par excellence in medical training and the program is responsible for ensuring that the recently graduated doctor reaches the established level of competence. A well-prepared assessment system with feedback is an effective tool to enhance the performance of future specialists and guarantee their qualification.

Objective:

To analyze the assessment system for pediatric residents at a university hospital, aiming to promote teaching training in assessment methods.

Methodology:

educational action research (research-teaching), conducted with teachers and preceptors in the pediatric residency of a university hospital. The stages consisted of: a) the application of a questionnaire on participant profile and assessment methods used to evaluate the medical residents; b) intervention (workshop) on the assessment of clinical skills and feedback; c) immediate assessment after the workshop, by applying another questionnaire based on level 1 of the Kirkpatrick model. Simple statistical analysis was used for the objective data and the content analysis, according to recommendations by Malheiros (2011) and Bardin (2013), for the qualitative part.

Result:

Ten (48%) of the 21 participants declared not being formally trained in assessment and that they applied the more traditional methods learned from personal experience. Regarding the assessment methods, 81% (17/21) of the participants used more than one, aiming to obtain a more encompassing and reliable assessment. However, none of the teachers/preceptors used a systematized assessment of the clinical skills or provided feedback to the medical residents. After the workshop, with a focus on assessing performance in a simulated ‘Objective Structured Clinical Examination’ (OSCE) environment, all the participants were favorable to apply the OSCE in their teaching-learning practice with residents and undergraduate students, reaching Kirkpatrick level 3.

Conclusion:

The action research helped identify limitations in the assessment and feedback system of the pediatric resident physician. The methodology used revealed an aggregating effect and contributed to the development of the collaborative and integrative sense in the group. However, it was not enough to positively interfere with the pediatric medical residency assessment in the short term.

Keywords: Medical Residency; Formative Feedback; Teacher Training; Performance Assessment

Resumo:

Introdução:

A residência médica é a especialização por excelência na formação do médico, e cabe ao programa assegurar que o residente egresso atinja o nível almejado de competência. Um sistema avaliativo bem elaborado e com feedback é ferramenta efetiva para aprimorar o desempenho do futuro especialista e garantir a qualificação dele.

Objetivo:

Este estudo teve como objetivo analisar o sistema de avaliação do médico residente em pediatria de um hospital universitário, com o intuito de promover a formação docente em métodos avaliativos.

Método:

Trata-se de uma pesquisa-ação educacional (pesquisa-ensino) realizada com docentes e preceptores da residência médica em pediatria de um hospital universitário. As etapas consistiram em: 1. aplicação de questionário sobre o perfil dos participantes e os métodos avaliativos utilizados com os residentes; 2. intervenção com a realização de um workshop sobre avaliação de habilidades clínicas e feedback; 3. avaliação imediata, após o workshop, com aplicação de outro questionário, elaborado com base no nível 1 do método Kirkpatrick. Utilizaram-se a análise estatística simples, para os dados objetivos, e a análise de conteúdo, segundo recomendações de Malheiros e Bardin, para a parte qualitativa.

Resultado:

Dos 21 participantes, dez (48%) informaram que não tinham capacitação formal em avaliação e que utilizavam métodos avaliativos mais tradicionais. Quanto aos métodos, 81% (17/21) dos participantes informaram que utilizavam mais de um, com finalidade somativa, para obter uma avaliação mais abrangente e fidedigna. No entanto, nenhum utilizava uma avaliação sistematizada de habilidades clínicas com fornecimento de feedback. Após o workshop com enfoque em avaliação de desempenho em ambiente simulado, em que se adotou o Objective Structured Clinical Examination (OSCE), os participantes utilizaram, com os internos, o método avaliativo do aprendizado na sua prática cotidiana, e, dessa forma, o treinamento atingiu o nível 3 de Kirkpatrick.

Conclusão:

A pesquisa-ação propiciou identificar limitações no sistema de avaliação e feedback do médico residente em pediatria. A metodologia utilizada revelou um efeito agregador e contribuiu para desenvolver o sentido colaborativo e integrativo no grupo. No entanto, não foi suficiente para interferir positivamente, em curto prazo, na avaliação da residência médica em pediatria.

Palavras-chave: Residência Médica; Feedback Formativo; Capacitação de Professores; Avaliação de Desempenho

INTRODUCTION

Medical residency is the specialization par excellence in the training of physicians, and the program is responsible for ensuring that recently graduated doctors reach the desired level of competence1. As it is in-service training, assessment, in this context, goes beyond the cognitive evaluation, constituting a daily challenge for the teacher and the preceptor.

An effective evaluation process requires, in addition to teacher training, a system that combines several types of evaluation, aiming to reach all elements of learning (knowledge, skills and attitudes), in addition to ensuring the validity and reliability of the utilized methods2.

Considering the complexity of the teacher’s role in medical education, the development and implementation of teacher development programs, activities that seek to improve the knowledge and skills of health professionals as teachers, should be considered a permanent process3),(4. However, according to Steinert4, despite the growing interest in the subject in recent years, few studies have focused specifically on teacher development for student evaluation.

The scenario of this research - our daily practice as teachers and preceptors in the field of pediatrics - is Hospital Universitário Professor Alberto Antunes (Hupaa) at Universidade Federal de Alagoas (Ufal), which has a medical residency program (MRP) in pediatrics that was implemented 30 years ago and has all the credentials to keep it effective to the present day. The program lasts three years, and five vacancies are made available annually. The occupancy rate of medical residency in pediatrics, from 2015 to the present day, was 93.3% (28/30), with a 100% rate of full residency training (18/18)5.

However, the evaluation process of the program’s residents still follow a traditional and summative characteristic, even though the National Medical Residency Commission (CNRM, Comissão Nacional de Residência Médica) has defined guidelines for the evaluation of resident doctors since CNRM Resolution N. 05, of November 12, 19796 (revoked), of which guidelines were reiterated in CNRM resolution N. 02, of May 17, 2006, considering the terms of article 137:

[...] Art. 13. In the periodic evaluation of the resident physician, the modalities of written, oral, practical or performance tests by an attitude scale will be used, which include attributes such as: ethical behavior, relationship with the health team and with the patient, interest in activities and others at the discretion of the institution’s COREME. §1. The minimum frequency of evaluations will be every three months. §2º. At the institution’s discretion, a monograph and/or presentation or publication of a scientific article may be required at the end of the training. §3º The criteria and the results of each evaluation must be made known to the resident doctor.

As for the evaluation, a conceptual model proposed by Miller, known for several decades, known as the Miller’s pyramid, demonstrated to teachers that, regarding professional development, the evaluation cannot be restricted to theoretical knowledge, as it is necessary for the student to know how to apply this knowledge, perform it, in a practical way, in simulated environments and, finally, apply it in real life8.

Miller’s pyramid aligns its strata with the educational objectives and evaluation methods aimed at the types of skills and competences whose domain one wants to know, rising from the theoretical knowledge contained at the base - “to know” and “to know how” - to “to show how” and “to perform”. The pyramid apex corresponds to the evaluation of professionals in their work environment9.

Based on these dimensions and the resident doctor’s degree of learning, the observations made by teachers and preceptors should be directed, in addition to the cognitive one, to the performance evaluations, considering the clinical and psychomotor skills, the interaction with the patient, the management of information, the capacity for judgment, synthesis and decision, as well as the preservation of ethical attitudes1.

Most clinical skills evaluation methods have, as a basic principle, the direct observation of the resident’s performance in clinical tasks, in a real or simulated environment. In this sense, the performance of feedback should be allowed, preferably an immediate (formative) one, which consists in describing and discussing with the residents their performance related to a given activity10.

A well-developed and periodic evaluation system, with continuous feedback, is an effective tool to improve the performance of the future specialists and guarantee their qualification, a goal of indisputable importance in the training process2. For that purpose, the resident’s evaluation needs the systematization and institutionalization related to how to evaluate, in addition to teacher training for this important aspect of the teaching-learning process.

This research proposed to answer the following question: how are residents being evaluated regarding the skills acquired in the pediatric medical residency program at Hospital Universitário Professor Alberto Antunes? Hence, the objective was to analyze the system used to evaluate the pediatric resident doctor of a university hospital, aiming to promote teacher training in evaluation methods.

METHODOLOGICAL TRAJECTORY

An educational action research was created aiming to identify gaps in pedagogical practice and cause changes in educational habits, considering its potential as an investigative praxis, in the resident’s evaluation process through an intentional sample with teachers and preceptors of the aforementioned residency program in pediatrics.

In the educational field, the action research (research-teaching) consists of an investigation about the practice itself and implies the awareness by the participants, allowing them to be involved in all phases of the methodological trajectory11),(12.

Its use in the educational field allows the researcher teacher to identify a problem in their pedagogical activity and, through research, create the conditions to transform it, aiming to favor the personal and professional growth of the researchers and the involved participants11),(12.

The action research follows a cycle in which practice is improved in the movement between acting in the field of the problem and investigating about it. The cycle includes the identification of the problem and data production on the effects of a change in practice during the intervention, before and after its implementation, using pre- and post-methods to monitor the effects caused by the change13) (Figure 1).

Figure 1 Diagram representing the development of the action research. 

According to Malheiros14, this methodology is very useful in the educational field, because it allows studies on changes in curriculum, teaching-learning models, evaluation methods, among other aspects.

In the research development, different procedures were used for data collection, divided into three stages, observing and complying with the action research methodology, aiming to understand the reality of the resident physician’s evaluation and to propose action strategies for its improvement.

In stage 1 of the action research cycle, all teachers and preceptors who worked at the pediatric medical residency in Hupaa were invited to participate in the research, comprising total of 22 individuals, of which 16 were teachers and 6 preceptors, from January 2018 to April 2019. Of the 22 participants, 16 were directly linked to the University Hospital and 6 to other health services (Hospital Geral do Estado Dr. Oswaldo Brandão Vilela and Hospital Escola Dr. Hélvio Auto), where the pediatric residents from Hupaa work in the urgency/emergency and infectious diseases sectors, respectively. In this phase, one researcher teacher was excluded from the study, and the others agreed to participate in the discussion and to answer the questionnaire.

The first semi-structured questionnaire was applied to identify the deficiencies in the residents’ evaluation process, consisting of questions related to the participants’ sociodemographic data, specific data on the teachers’ training in assessment, the evaluation methods (cognitive tests or assessment of clinical skills) that they used with the resident physicians, the factual knowledge of performance evaluation methods in real and simulated environments, the perspective of teachers and preceptors regarding the evaluation process, as well as the intention to participate in a teaching development workshop about an evaluation method.

The participants’ answers to the open questions were organized based on the ideas that emerged from the guiding questions, when the pre-analysis was carried out through further reading, observing the emergence of categories that were not previously created. A matrix was created, and all statements were transcribed in full. The participants were coded by letters and numbers, following the order of analysis of the questionnaires - teacher (T) and preceptor (P).

Matrices were created, which stored the explicit or implicit ideas, the creation of the categories and the registration units that associate the statements to the topic to explain, in the text, how the result was achieved. The focal points and recording units were interpreted, and the synthesis for each focus was developed14.

Simple statistical analysis was used for objective data and content analysis, according to the recommendations by Malheiros14 and Bardin15, for the qualitative part.

Stage 2, an action resulting from the search for a solution to the diagnosed problem, led to the planning of a pedagogical intervention. According to Malheiros14, in this model of action research, an intervention is made in a certain reality so that, subsequently, its results can be evaluated.

This step consisted in planning the intervention phase, after analyzing the responses to the questionnaire. Then the training workshop on clinical skills assessment was developed for teachers and preceptors, focusing on the evaluation method chosen by the participants.

Aiming to maintain the integrative strategy, the dates for the training were reached by consensus with all involved participants (teachers, preceptors, actors and pediatric monitors). The entire workshop process was recorded through photos and footage, with the participants’ consent.

The stage 3 of this cycle, three weeks after the workshop, was characterized by the evaluation of the pedagogical intervention through a semi-structured questionnaire, sent by electronic mail, based on the Kirkpatrick model, which consists in an internationally recognized evaluation model, of which purpose is assessing educational actions aimed at professionals. It consists of four levels of training evaluation: (1) reaction, (2) learning, (3) behavior (transference) and (4) results16.

The questionnaire was designed based on the level 1 of the Kirkpatrick model. At this level, the participants’ reaction to the training itself is evaluated, as well as their reactions to the learning experience. Elements were questioned, such as the program content, the achieved expectations, the assessment of infrastructure and logistics (facilities and equipment), the duration and organization of training, the quality and content of the didactic material, the structure of scenarios, as well as the speakers’ evaluation (didactics, communication, interaction and knowledge) and the methodology used. Moreover, there were open questions about the motivation to participate in the workshop and the willingness to use the OSCE method in teaching practice with the resident physicians.

Based on the action research cycle, the trajectories used are described in the flowchart below, with the research strategies and constructed data analysis techniques (Figure 2).

Figure 2 Research flowchart. 

The project was approved by the Research Ethics Committee (REC) of Universidade Federal de Alagoas, under Opinion n. 2,304,092 (CAAE: 74854717.0.0000.5013), with no conflicts of interest.

RESULTS AND DISCUSSION

Stage 1 - Situational diagnosis

At this stage, regarding the diagnosis of the problem, all teachers and preceptors who worked in the pediatric medical residency and who met the inclusion criteria participated in the study, totaling 21 individuals, of which 15 were teachers and 6 were preceptors. Of these, 76% (16/21) were females, whose age varied between 29 and 62 years (mean age of 46.5 years), training time from 7 to 38 years (mean of 22.3 years) and teaching time between 1 and 37 years (mean of 13.6 years). Regarding the degree, 23.8% (5/21) have a Doctorate Degree and 33.3% have a Master’s Degree (7/21).

Of the 21 participants, 10 (48%) reported they had no formal training in evaluation and that they used traditional assessment methods, based on their own experience in the service environment. This percentage includes all preceptors, which is due to the fact that there is no requirement for pedagogical training to work in preceptorship17, although it is the preceptor’s role to teach, monitor the daily practice and evaluate physicians in training. Preceptors, in addition to good specific knowledge in the field, need teacher development activities and permanent institutional support to help them improve their teaching skills18.

As for the evaluation methods, 81% (17/21) of the participants reported using more than one, for summation purposes, to obtain a more comprehensive and reliable evaluation. According to Norcini et al.19, the structure for a good evaluation consists of an organized combination of methods, to constitute an evaluation system; however, none of the teachers/preceptors uses a systematized evaluation of the resident physician’s clinical, psychomotor or affective skills, or provide feedback. Therefore, despite the diversity of evaluation methods, there is no guarantee of the absence of weaknesses20.

In the statements provided by teachers and preceptors, a collective concern with this inadequacy related to the evaluation methods can be observed:

Great research. I was always bothered by the evaluation without resources. I believe that there are risks of injustice. (T2)

I recognize how poor the traditional methods of assessment are. Perhaps due to accommodation and resistance to changes, we have not yet managed to leave this comfort zone. Open to new experiences. (T6)

This corroborates the results of Zimmerman et al.20, where teachers declared having difficulty in evaluating and formulating tests due to the lack of theoretical basis, in addition to the lack of standardization in the medical course.

The importance of performance evaluation, at this level of training, is due to its potential to verify clinical skills (communication, physical examination and procedures), and the continuous and formative evaluation allows the correction of failures and reduces the possibility of errors.

Ross et al.21, in a retrospective cohort study, analyzed the performance and progression of resident physicians when assessed by a traditional summative evaluation system, compared to a competency-based evaluation system. From this perspective, they showed the effectiveness of approaching residents that have difficulties, focusing on the possible correction of failures. The traditional evaluation methods identified the residents’ problems, but were not effective in correcting the existing gaps, perhaps because evaluations were disconnected from daily observations.

In this research, it was observed that the most frequently used method is the global evaluation, with 76.2% (16/21) of the answers, followed by the observation of the student in an actual environment, used by 43% (9/21) of the participants. As for the evaluation in a simulated environment, two participants (9.5%) reported having had some training for this model, although they did not use it with the residents. The other instruments mentioned were related to cognitive evaluation methods, such as discussion of clinical cases (10/21 or 47.6%), multiple choice tests (7/21 or 33.3%) and oral evaluation (7/21 or 33.3%) (Chart 1).

Chart 1 Distribution of the evaluation methods used by teachers and preceptors of the pediatric medical residency at the UFAL University Hospital, 2018. 

The global evaluation consists of a scale that assesses knowledge, punctuality and attitude. According to literature data, which corroborate the findings of this article, the global evaluation is the most frequently used instrument by all postgraduate courses for the evaluation of skills in the United States22, for instance. However, the authors observed two major limitations of the global evaluation. The most important of them is related to the fact that a physician with a deficiency in certain area can achieve a satisfactory global classification, if they show a good result in another competence. The other limitation is that this instrument provides little or no information for constructive feedback, an important component for the development of resident physicians23.

To validate the global concept, the proposal is to construct a matrix with specific items, which reflect the combination of the necessary attributes for good professional performance, in addition to the fact that the resident is being evaluated by several teachers23.

The observation of the student in a real-life environment is a usual method of evaluation, of easy operationalization, favorable to the provision of immediate feedback and widely accepted by the participants2. However, the absence of systematization, the use of an evaluation form (checklist) or providing feedback, generates a low degree of reliability.

The evaluation in a simulated environment allows the approach of clinical tasks that represent what is common in medical practice, at a level of difficulty compatible with the stage of training, and to ensure that all students are evaluated under similar conditions, with careful observation and with the use of the checklist2. However, this method is not used by any of the participants with the residents, although two of them are aware of it. This fact may suggest the absence of an established evaluation process, the lack of interaction between the participants and the difficulty for its operationalization, as this method requires teamwork, prolonged performance and more costly logistics (scenarios, actors, evaluation forms).

Using evaluation methods for clinical skills, whether in real or simulated environments, is not an easy task, as it involves choosing the most appropriate method that relies on a valid, reliable, viable and acceptable technique for everyone involved in the process2. It also implies teacher training in evaluation, teamwork, participation in action research and other strategies, in addition to standardization and systematization. Some participants expressed themselves positively in relation to the research and the need to standardize evaluation:

I hope that after this research and the obtained results, there is a standardization of the evaluation methods used in FAMED. It is very important to standardize the assessments. (T1)

(I hope) It becomes a project, registered with the teaching management, including preceptors from all clinics. (T4)

In this sense, 90% of the participants mentioned the need for training in evaluation methods and showed interest in participating in training about a specific evaluation method in a simulated environment, as explained in the statement by P2: “I am interested in teaching trainings”. Or, even, when one addresses the need for a performance evaluation method such as the OSCE, for instance: “Include OSCE as an integrated assessment during the pediatric internship in the 9th and 10th semesters (at least 1x/semester)” (T5).

The evaluation improvement of teachers and preceptors ensures the quality of the evaluation and the teaching-learning process, since the evaluation allows the review of educational planning and adjustments in their teaching practice20.

One of the statements refers to the need for feedback, as stated by T3: “Continuous assessment, in daily life, is essential for the learning process, always with feedback for its strengthening”. In this statement, it can be observed that the teacher synthesizes the entire process of one evaluation with a formative purpose.

Feedback is the substrate of a formative assessment and an effective tool to improve student performance, especially when it is performed immediately, after the clinical task10. Therefore, it must take place in a dialogical way, with the student playing an important role in the evaluation of their own performance24.

In medical education, feedback is as essential for educators to promote learning as it is for students, as it provides information about their work and quality, aiming to generate improvements25.

The results showed that the participants’ concern about the residents’ assessment methods was a collective one, considering this as the first step towards institutionalizing changes within the teaching-learning scenario.

Stage 2 - Workshop: integrative and interactive dynamics

After identifying the problem, evidenced by the absence of systematization, institutionalization and teacher training regarding evaluation methods for residents in pediatrics, an intervention was planned and implemented as a teaching development strategy under evaluation. Most participants chose the Objective Structured Clinical Examination (OSCE) training.

The OSCE is an assessment tool in a simulated environment, situated in the “show how you do it” dimension of Miller’s pyramid. It is a very frequently used and appreciated evaluation method in several parts of the world. When well prepared, it brings important information for the future professional’s performance.

Nine (43%) research participants attended the workshop with simulated activity: 7 teachers and 2 preceptors, just under 50% of the total sample, despite the previously agreed schedules. Observers and guests also attended the workshop, such as pediatric residency managers, resident physicians and monitors of the pediatric discipline. The activity was carried out at the School of Medicine of Universidade Federal de Alagoas, located next to Professor Alberto Antunes University Hospital. As described by Steinert4, the proximity to the workplace is one of the facilitating factors to increase the participation, motivation and access to these educators. Of those present participants, 8 were teachers and preceptors linked to Hupaa. It was observed that only one of the six participants was linked to the other institutions. Justifications about the absence were conveyed.

The workshop consisted of two sequential moments: in the first, the researchers exposed general data on evaluation and on the OSCE method (in the classroom); the second part was directed to the workshop practice, held in the tutoring rooms of the School of Medicine, with the actors and dummies from the Skills Laboratory.

At that time, the research participants were instructed to rotate through the different scenarios, in groups of three or four. It is noteworthy that the moving through the stations, in the OSCE, is done by the assessed student; however, on this occasion, the objective was not to evaluate the resident, but to allow the participating teacher and preceptor to have an experience with different scenarios, evaluation forms and feedback.

At the end of the circuit, the participants returned to the classroom, where the phases of the OSCE that were carried out were presented, with photographic records of the workshop.

Stage 3 - Evaluation of the impact of the workshop

The Kirkpatrick method was used in this stage, aiming to assess the outcome of the OSCE workshop. Thus, level 1 of the Kirkpatrick method (reaction), used in the evaluation of the training, measures the participants’ impressions in relation to the content, the instructors, the materials and resources, the environment and the facilities. According to Kirkpatrick, all programs must be evaluated at this level, aiming to promote improvements. A positive reaction does not necessarily guarantee learning, but a negative reaction and dissatisfaction certainly reduce the possibilities for learning26),(27. In this respect, the 12 items of the evaluation obtained a predominance of positive assessments, with 10 items being evaluated as between excellent and good.

Regarding the item related to the application of the evaluation method in training (OSCE), all participants were favorable to using it in their teaching-learning practice with resident physicians and also with undergraduate students. This corroborates the comments on the motivation to attend the workshop, when they refer, basically, to the teaching development in more innovative evaluation methods:

New learning. (T1)

Evaluation of students. (T3)

Improve the evaluation process. (T6)

Improve my performance with residents and students. (P1)

In the activity, it was also possible to include the participation of master’s degree students, the professional master’s degree in health teaching, the discipline of teaching evaluation, as a form of learning through the observation of the method being performed in real time.

It is emphasized that a positive action, in a team, undergoes an idealization and generates a new action, according to the cycles of educational action research. It is also noteworthy that the training reached Kirkpatrick level 3, since the participants used the appraisal method of learning in their daily practice, now together with the interns. The activity took place under the same structure of the workshop, which corroborates Steinert’s4 thinking, when she states that bringing teacher development activities into the workplace increases participation, motivation, access and, that together with individual engagement, a sequencing of activities and continued guidance of participants, ensures better training effectiveness.

The butterfly effect28 caused by the intervention (workshop), in the educational action research proposal, and through the evaluation method itself, has also become a teaching-learning environment for students from other levels of education. It is expected that this effect, a metaphor used in science, described by Edward Lorenz in 197228, caused by a small change at the beginning of an event, can have positive future consequences and that it will unfold into new interventions in evaluation methods at the institution, for all undergraduate and graduate levels. In this sense, it is in education and in the reflections that come from it that the transformative potential of changes ensues.

The limitations of the study include the difficulty in aggregating all the participants involved (teachers, preceptors and actors), due to the conciliation of time, place and work, in addition to addressing only one method of evaluation, due to the organization’s logistics and practicality.

FINAL CONSIDERATIONS

The action research carried out with the teachers and preceptors who work with the medical residency program allowed, mainly, to identify limitations in the evaluation and feedback system of the pediatric resident physician, such as the inadequacy of the methods used and the perception that teachers and preceptors should be more aligned with a formative assessment and feedback. In this sense, it should cover all elements of teaching-learning of practice in service, such as clinical skills, attitude, ethics, clinical reasoning and professionalism, thus allowing the identification and correction of any flaws of the physician in-training.

The methodology used in the research disclosed an aggregating effect and contributed to the development of a collaborative and integrative sense within the group, awakening, through protagonism, the need for teacher training. Based on that, some positive changes were observed in the participants (teachers and preceptors) regarding the assessment at other levels of teaching-learning (undergraduate school and internship).

Although reflections on the inadequacy of the evaluation process have been generated, the intervention (workshop) was not sufficient to positively interfere, in the short term, with the pediatric medical residency evaluation.

It is known that a well-designed evaluation process requires time, periodicity, planning and organization. Solid changes require other factors, in addition to continuing teacher training in teaching-learning; it requires dedication, commitment from the involved educators, teachers and preceptors, student awareness and a de facto institutional formalization.

Evaluating is a complex process that remains a challenge; however, promoting spaces for debates and reflections, as shown by the research, can lead to significant measures and they might evolve towards the educational objectives aimed at the physician in training.

REFERENCES

1. Megale L, Gontijo ED, Motta JAC. Avaliação de competência clínica em estudantes de medicina pelo Miniexercício Clínico Avaliativo (Miniex). Rev Bras Educ Med. 2009;33(2):166-75. [ Links ]

2. Troncon LEA. Avaliação de habilidades clínicas: os métodos tradicionais e o modelo “OSCE”. Olho Mágico. 2001;1:8-12. [ Links ]

3. Perim GL, Abdalla IG, Aguilar-da-Silva RH, Lampert JB, Stella RCR, Costa NMSC. Desenvolvimento docente e formação médica. Rev Bras Educ Med . 2009;33(1):70-82. [ Links ]

4. Steinert Y. Faculty development: from rubies to oak. Med Teach. 2019;41, p. 429 - 435. doi: 10.1080/0142159X.2019.1688769. [ Links ]

5. Hospital Universitário Professor Alberto Antunes. Comissão de Residência Médica. Relatório. Maceió: Hupaa; abr. 2020. [ Links ]

6. Brasil. Resolução CNRM nº 05 de 12 de novembro de 1979. Estabelece especialidades médicas credenciáveis como Programa de Residência Médica e dá providências adicionais. Brasília: Ministério da Saúde, Comissão Nacional de Residência Médica; 1979. [ Links ]

7. Brasil. Resolução CNRM nº 02, de 17 de maio de 2006. Dispõe sobre requisitos mínimos dos Programas de Residência Médica e dá outras providências. Brasília: Ministério da Saúde, Comissão Nacional de Residência Médica ; 2006 [access in 23 jun 2019]. Available from: Available from: http://portal.mec.gov.br/index.php?option=com_docman&view=download&alias=100951-resolucao-cnrm-n-2-de-17-de-maio-de-2006&category_slug=novembro-2018-pdf&Itemid=30192 . [ Links ]

8. Souza RGS. Atributos fundamentais dos procedimentos de avaliação. In: Tibério IFLC, Daud-Gallotti RM, Troncon LEA, Martins MA, organizadores. Avaliação prática de habilidades clínicas em medicina. São Paulo: Atheneu; 2012. p. 1-11. [ Links ]

9. Panúncio-Pinto MP, Troncon LEA. Avaliação do estudante: aspectos gerais. Medicina (Ribeirão Preto). 2014;47(3):314-23. [ Links ]

10. Zeferino AMB, Passeri SMRR. Avaliação da aprendizagem do estudante. Cadernos ABEM. 2007;3:39-43. [ Links ]

11. Zaidan S, Ferreira MCC, Kawasaki TF. A pesquisa da própria prática no mestrado profissional. Revista Plurais. 2018;3(1):88-103. [ Links ]

12. Thiollent M. Metodologia da pesquisa-ação. São Paulo: Cortez; 2002. [ Links ]

13. Tripp D. Pesquisa-ação: uma introdução metodológica. Educação e Pesquisa. 2005;31(3):443-66. [ Links ]

14. Malheiros BT. Metodologia da pesquisa em educação. Rio de Janeiro: LTC; 2011. [ Links ]

15. Bardin L. L’analyse de contenu. 2ème ed. Paris: Presses Universitaires France; 2013. [ Links ]

16. Kirkpatrick J, Kirkpatrick WK. The Kirkpatrick four levels: a fresh look 50 years 1959-2009. Kirkpatrick Partners; 2009 [access in 1º nov 2019]. Available from: Available from: http://www.kirkpatrickpartners.com/Portals/0/Resources/Kirkpatrick%20Four%20Levels%20white%20paper.pdf . [ Links ]

17. Botti SHO. O papel do preceptor na formação de médicos residentes: um estudo de residências em especialidades clínicas de um hospital de ensino [tese]. Rio de Janeiro: Escola Nacional de Saúde Pública Sergio Arouca; 2009. [ Links ]

18. Almeida MTC, Batista NA. Ser docente em métodos ativos de ensino-aprendizagem na formação do médico. Rev Bras Educ Med . 2011;35(4):468-76. [ Links ]

19. Norcini J, Anderson MB, Bollela V, Burch V, Costa MJ, Duvivier R, et al. Consensus framework for good assessment. Med Teach . 2018; 40(11):1102-9. [ Links ]

20. Zimmermann MH, Silveira RMCF, Gomes RZ. O professor e a arte de avaliar no ensino médico de uma universidade no Brasil. Rev Bras Educ Med . 2019;43(3), p. 5-15. [ Links ]

21. Ross S, Binczyk NM, Hamza DM, Schipper S, Humphries P, Nichols D, et al. Association of a competency-based assessment system with identification of and support for medical residents in difficulty. JAMA Netw Open. 2018;1(7):1-12 [access in 30 nov 2019]. Available from: Available from: https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2713042 . [ Links ]

22. Silber GC, Nasca TJ, Paskin DL, Eiger G, Robeson M, Veloski JJ. Do global rating forms enable program directors to assess the ACGME competencies? Acad Med. 2004;79(6):549-56. [ Links ]

23. Domingues RCL, Amaral E, Bicudo-Zeferino AM. Conceito global: um método de avaliação de competência clínica. Rev Bras Educ Med . 2009;33(1):148-51. [ Links ]

24. Ramani S, Krackov SK. Twelve tips for giving feedback effectively in the clinical environment. Med Teach . 2012;34:787-91. [ Links ]

25. Qureshi NS. Giving effective feedback in medical education. Obstet Gynaeco. 2017;19:243-8. doi: 10.1111/tog.12391. [ Links ]

26. Mira VL, Peduzzi M, Melleiro MM, Tronchin DMR, Prado MFF, Santos PT, et al. Análise do processo de avaliação da aprendizagem de ações educativas de profissionais de enfermagem. Rev Esc Enferm USP. 2011;45:1574-81 [access in 19 sep 2019]. Available from: Available from: http://www.scielo.br/scielo.php?script=sci_arttext&pid=S0080-62342011000700006&lng=en&nrm=iso . [ Links ]

27. Silva NP, Francisco AC, Hatakeyama K, Silva MCG. Avaliando as práticas de educação corporativa à distância por meio do modelo de Kirkpatrick: um estudo de caso numa empresa do ramo de energia no estado do Paraná. Emancipação. 2010;10(2):501-15. [ Links ]

28. Ferrari PC, Angotti JAP, Tragtenberg MHR. Educação problematizadora a distância para a inserção de temas contemporâneos na formação docente: uma introdução à Teoria do Caos. Ciênc Educ (Bauru). 2009;15(1):85-104. [ Links ]

2Evaluated by double blind review process.

SOURCES OF FUNDING The authors declare no sources of funding.

Received: December 27, 2019; Accepted: April 19, 2021

brunaduarte100@gmail.com camposdelisboa@gmail.com analydiavet@hotmail.com

Chief Editor: Daniela Chiesa. Associate Editor: Daniela Chiesa.

AUTHORS’ CONTRIBUTION

Bruna de Sá Duarte Auto: data collection, analysis, interpretation, and discussion; writing and review of the manuscript. Maria Viviane Lisboa de Vasconcelos: research orientation; data analysis, interpretation, and discussion; writing and review of the manuscript. Ana Lydia Vasco de Albuquerque Peixoto: data discussion and review of the manuscript.

CONFLICTS OF INTEREST

The authors declare no conflicts of interest.

Creative Commons License This is an open-access article distributed under the terms of the Creative Commons Attribution License