SciELO - Scientific Electronic Library Online

 
vol.23JOINT CONSTRUCTION OF PRACTICES AS A CURRICULUM COMPONENT BY EDUCATORS OF TRAINING CHEMISTRY TEACHERS: AN ANALYSIS BASED ON THE THEORY OF COMMUNICATIVE ACTINGMOTIVATION TO STUDY PHYSICS author indexsubject indexarticles search
Home Pagealphabetic serial listing  

Services on Demand

Journal

Article

Share


Ensaio Pesquisa em Educação em Ciências

Print version ISSN 1415-2150On-line version ISSN 1983-2117

Ens. Pesqui. Educ. Ciênc. vol.23  Belo Horizonte  2021  Epub Aug 17, 2021

https://doi.org/10.1590/1983-21172021230117 

ARTICLE

STUDENT ENGAGEMENT IN A REMOTE AND EMERGENCY PHYSICS COURSE

Helder de Figueiredo e Paula1 
http://orcid.org/0000-0002-3860-8009

Sérgio Luiz Talim1 
http://orcid.org/0000-0003-1650-7327

Cecília Siman Salema2 
http://orcid.org/0000-0002-9725-4466

Vinícius Reis Camillo2 
http://orcid.org/0000-0002-5581-9940

1Universidade Federal de Minas Gerais, Colégio Técnico, Setor de Física, Belo Horizonte, MG, Brasil.

2Universidade Federal de Minas Gerais, Instituto de Ciências Exatas, Belo Horizonte, MG, Brasil.


ABSTRACT:

In this paper, we show evidences of student engagement in an introductory, remote and emergency Physics course. This experience occurred during the COVID-19 pandemic, in a high school that provides professional education. We present the pedagogical challenges faced by Physics teachers and justify their choice to evaluate student engagement. Our data stemmed from a questionnaire answered by students after the end of the first module of the course. We identified three engagement dimensions: cognitive-behavioral, emotional and social. Our data show that there was engagement, predominantly, in the cognitive-behavioral and emotional dimensions.

Keywords: Engagement; Emergency Remote Education and COVID-19; Science Teaching

RESUMO:

Neste trabalho, mostramos evidências de engajamento de estudantes que participaram de um curso introdutório, remoto e emergencial de Física. O curso foi realizado durante a pandemia da COVID-19, em uma escola que oferece ensino técnico integrado e profissionalizante de nível médio. Apresentamos os desafios pedagógicos enfrentados pelos professores de Física e justificamos a escolha desses sujeitos por avaliar o engajamento de seus estudantes. Os dados da pesquisa emergiram de um questionário que foi respondido pelos estudantes após o término do primeiro módulo do curso. A utilização do questionário nos permitiu identificar três dimensões do engajamento: cognitivo-comportamental, emocional e social. Nossos dados mostram que houve engajamento das/dos estudantes, predominantemente, nas dimensões cognitivo-comportamental e emocional.

Palavra-chave: Engajamento; Ensino Remoto Emergencial e COVID-19; Ensino de Ciências

RESUMEN:

En este artículo mostramos evidencia del compromiso de los estudiantes que participaron en un curso introductorio, remoto y de emergencia de Física, realizado durante la pandemia COVID-19, en una escuela secundaria que ofrece educación profesional. Presentamos los desafíos pedagógicos que enfrentan los profesores de física y justificamos su elección para evaluar el compromiso de sus estudiantes. Los datos de la investigación surgieron de un cuestionario que fue respondido por los estudiantes una vez finalizado el primer módulo del curso. Identificamos tres dimensiones de lo envolvimiento de los estudiantes: cognitivo-conductual, emocional y social. Nuestros datos muestran que los estudiantes se involucraron, predominantemente, en la dimensión cognitivo-conductual y la dimensión emocional.

Palabras clave: Participación de los estudiantes; Educación Remota y COVID-19; Enseñanza de las Ciencias

Introduction

The research hereby disclosed was carried out in the capital of a Brazilian state. In this city, by decision of the Municipal Government, in-person classes were suspended in mid-March 2020, in all educational institutions. This measure complied with the World Health Organization sanitary protocols to reduce the spread of the COVID-19 pandemic.

The subjects of our research are students from a high school that provides professional education. When this text was written, in-person classes remained suspended. To designate this reality, we make use of the expression Emergency Remote Teaching (ERT) instead of Distance Education (DE), as proposed by Arruda (2020), as well as by Saraiva, Traversini and Lockmann (2020).

On the one hand, Arruda (idem) differentiates DE from ERT in order to: (1) criticize what he considers to be a prejudiced idea about DE, which stems from the misconception that this teaching modality is inferior to traditional education; (2) highlight the precarious quality of many ERT experiences, which differ from DE due to the absence of the required infrastructure or subjects with proper training.

On the other hand, Saraiva, Traversini and Lockmann (idem) differentiate DE from ERT based on the different rationalities in which these teaching methods/modalities are grounded. The former is based on a goal-oriented rationality, with no regulation of time or space where students execute tasks. The latter inherited the disciplinary rationality, which is characteristic of traditional education and, this way, transfers the responsibility from the teachers to the families, who are now obliged to exert disciplinary control over the students’ behavior.

In our framework, ERT started in the school in August of 2020, five months after classes had been suspended. Some public school students were unable to access the online environment due to social inequality and the absence of public policies regarding digital inclusion. The school under our investigation assisted the students who could not pay for Internet service. Nevertheless, part of the students accessed the online activities from smartphones. Subsequently, the school rented laptops and lent them to those students.

An almost entirely new course had to be conceived due to the establishment of ERT. The school’s Physics teachers intended to evoke, evaluate, and sustain the students’ engagement. Such fact could be acknowledged because one of the authors of this paper is part of this group of teachers. These teachers prepared a self-administered and anonymous questionnaire in which students were invited to evaluate the first module of the course, in addition to their own engagement in the ERT teaching and learning activities.

Our research group decided to analyze the data collected from the questionnaire in order to answer the following questions: (a) was there student engagement with the course? (b) if so, how to characterize this engagement?

Besides this brief introductory section, this paper consists of four other sections. In Section II, there is a description of the characteristics of the school and the Physics education provided to students before and during the pandemic; in Section III, we present the research theoretical framework and methodology; in Section IV, there is a description of the processes behind the construction of our data; in Section V, we present a general discussion of the results and research limitations, as well as the possible contributions to other teachers and researchers.

Research context

This research took place in a federal public school. Students join the school through a selection process.14 The experience of two of the co-authors of this paper in class councils shows that teachers’ assessment of the students’ involvement in different subjects is considerably varied. The performance of the students on the proposed activities, as well as their grades, also varies significantly depending on the subject. These findings indicate that student engagement in the school depends on the school subject.

In the Physics subject, most students attain satisfactory levels of academic performance, even though many complex topics are included in the course.2 We have carried out studies in which students were asked to evaluate the first grade of the introductory Physics course in the school (Paula and Talim, 2012, 2015). The strategy of having students rate the Physics course is a recurring practice in the school. Successive ratings indicate that most students are interested in the course topics.

This history with several signs of student engagement in the Physics course did not give the team of teachers good expectations in relation to the ERE, because the course underwent some profound changes. In our subject school, students have four fifty-minute Physics classes a week, grouped in sets of two classes. In one of the classes from the set, all students are gathered in a conventional classroom. In the other, students are gathered in a lab where they form small groups. These groups sit at rectangular tables.

In both settings, students often gather in groups and handle experimental materials. All the concepts, models and theories addressed in the course are presented as resources to the interpretation of experimental results, technological devices or phenomena present in their daily lives. Mathematical formalism is employed insofar as it contributes to these interpretations. Furthermore, there is a great rhetorical investment in the proposal and justification of problems on which each activity is based. This orientation is coherent with the epistemology developed by Bachelard (2001, p. 166), who states that all knowledge is the answer to a question/problem, and that if there is no problem, there cannot be scientific knowledge.

In the course, there was a great deal of interaction between the students and the Physics teacher, and also between the students and their peers. The interpretation of the phenomena discussed throughout the course is done in a collaborative manner, through discursive interaction between teacher and students, as well as between students and their peers. Activities assigned to be done at home are rare and structured so that they may be done in little time. Hence, it is fair to say that the course is almost completely face-to-face and predominantly experimental.

Because of ERT, activities that had been refined over the years proved to be useless. Digital apps with simulations of physical phenomena, which had previously been auxiliary resources in the “real lab”, turned into the predominant mediational resources of the proposed activities. Furthermore, ERT eliminated the possibility of a synchronous space where meaning could be negotiated through interaction between students and teachers, because the pandemic and the available conditions for the continuity of the pedagogic work demanded a predominantly asynchronous setting.3

This kind of ERT hampered the hearing and reception of the students’ contributions to the interpretation of real events and phenomena (Freitas e Aguiar Júnior, 2012), as well as the use of questions to guide them through the scientific concepts, models, and theories (Chin, 2007). Since students spend half of the course load inside the labs where they worked in small groups, peer interaction - whose importance to learning has already been discussed by many researchers (Driver et al., 1999; Barbosa e Jófili, 2004) - was also restricted.

When ERT started, some teachers and school staff consulted all the teachers and decided to divide the school subjects into two groups. The subjects were organized in modules. For most teachers, there is now an interval of at least four weeks between two consecutive modules of interaction with students.

Before the pandemic, the school had already made an arrangement with Google and started to use the GSuite platform.4 At the beginning of the pandemic, the student council did a survey among the students and claimed for the utilization of a single platform for the Emergency Remote Teaching. Students predicted that they would have difficulty learning how to use different platforms. The school’s previous agreement with Google imposed an intensive use of the GSuite apps.5

Subject to this reality, the Physics teachers had to conceive a course fully held on this platform. For each topic of the Physics ERT, an assignment was designed on Google Forms and a chat room was created on Google Chat. The chat rooms associated with each assignment were the main channel of asynchronous interaction between teachers and students. Synchronous interaction was restricted to weekly online meetings through Google Meet. Attendance in these meetings or watching their recordings, however, was not mandatory. The attendance rate in the synchronous meetings was low. For this reason, we believe that the students’ course evaluation - which formed the research data - was focused on the asynchronous activities.

As far as possible, the online activities that were structured as forms followed some of the basic guiding principles from the Physics pedagogical project. The progressive introduction of Physics concepts was done through videos that presented either real experiments or explorations in virtual labs. In this manner, the concepts continued being used as resources to the interpretation of experimental results and daily situations. Furthermore, students continued being requested to make predictions, formulate hypotheses, and articulate explanations to physical phenomena to which they are exposed.

Theoretical framework and research methodology

All Physics teachers in the school hold a PhD in Science Education and have research experience. One of the first grade teachers, who did not participate in this research, investigated student engagement in his former research and used Fredricks et al. (2004) as one of his main references.6 This is relevant because it allows us to understand the choice of this team of teachers towards developing a questionnaire to evaluate student engagement in the course.

The research group is composed of one of the first grade Physics teachers, a retired teacher from the same school, and two undergraduate students. The research group proposed modifications on the first version of the questionnaire made by the first grade Physics teachers. The questionnaire is, therefore, a result of cooperation between the research group and the teacher team.

The data analysis and discussion happened on online meetings with the members of the research group. The undergraduate students attended all meetings and were partially responsible for the literature review. Such review could not be presented in this article because the text reached the maximum size allowed by this journal and we preferred to favor the exposition of the data and the details of our analysis procedures in the available space. The second co-author conceived the architecture of the statistical analysis. The first coauthor advised the undergraduate research students, was partially responsible for the questionnaire since its first version,7 and was the main writer of this paper.

The fact that the data were collected from a questionnaire conceived and used by teachers as part of their teaching strategy led us to believe that submitting this research to an ethics committee was not necessary, although this would have been ideal. Since we committed to present the data and analysis to the teacher team in order to make eventual changes in ERT that could evoke or sustain student engagement, our research schedule was tight. Because of this agreement between the teacher team and the research group, the research had to be done within a timespan of four weeks.

We decided to carry out the research without formal consent of an ethics committee because we took into consideration the fact that our data were not collected solely for the research, and that the questionnaire was optional and anonymous. Furthermore, the teacher team authorized us to do the research.

Student engagement with the school and with specific school subjects became an eminent object of research throughout the 1990s and the beginning of the 2000s. In spite of the great variety of conceptions of engagement (Christenson, Reschly e Wylie, 2012), there is a certain consensus that it is a predictor of school achievement. Similarly, disengagement is considered to be a predictor of school dropout and it is associated with a series of behaviors and attitudes that precedes it, such as poor attendance, poor performance in tests and exams, apathy, and disruptive behavior (Klem & Connell, 2004).

Another general agreement found in the literature (Christenson, Reschly e Wylie, 2012; Fredricks et al., 2004; Fredricks et al., 2016) is the affirmation that the concept of engagement is multidimensional and that it involves several factors (emotional, cognitive, behavioral, attitudinal, social, among others), though they can vary in number and in the indicators of each dimension. In our research, we adopted the concept of engagement developed by Fredricks et al. (2016), for whom it involves four irreducible, interdependent and inseparable dimensions: behavioral, cognitive, emotional and social. It is important to clarify that, although the concept of engagement involves several dimensions - according to different authors - it should be interpreted as a single construct.

According to Fredricks et al. (2004), behavioral engagement refers to students’ compliance with school rules and absence of disruptive behavior, as well as participation, attention, persistence, and effort in the school activities. Emotional engagement concerns the affective reactions of students to teachers and activities, as well as the feeling of belonging and identification with the school. Cognitive engagement is related to students’ investment in learning, which involves metacognitive strategies to solve problems and understand complex ideas. The social dimension of engagement, as described by Fredricks et al. (2016), refers to the quality of the social interactions that are part of the learning and teaching environment.

Considering our collaboration with the group of teachers that created the questionnaire from which we collected our data, we suggested the use of the concept of engagement described by Fredricks et al. (2016). Besides these authors’ work, we found in the literature several other instruments for measuring engagement via self-administered questionnaires (Wang et al., 2016; Gale et al., 2015; Wang et al., 2014; Wang et al., 2019; Veiga, 2016). These instruments, however, were designed to be used in face-to-face courses. We found a small number of works that involved Distance Education (Nasir et al., 2020; Dixson, 2012 and 2015).

Many of these authors carried out pilot studies with samples of students as a validation step to their research instrument. Subsequently, these researchers interviewed the students in focal groups in order to understand their interpretation of the questionnaire content. We did not apply these validation steps to our instrument. Firstly, because it was not conceived solely for our research. Secondly, because we did not have the time or the necessary conditions to do so. For these reasons, we did not completely validate our research instrument, although we did carry out satisfactory procedures that allowed us to answer our research questions, as described in the next section of this paper.

Our questionnaire was inserted on a form entitled “Evaluation of the Module about Waves”. In its introduction, all the three Physics teachers who were responsible for the course confessed the challenge of transforming a course that was previously 100% face-to-face and predominantly experimental into a course with asynchronous and online activities. In the same introduction, they also committed to improve students’ experiences during the next ERT module. Finally, the introduction also informed students that all collected data would be anonymous and that answering the questionnaire was optional, although of utmost importance to help enhance the course.

Following the introduction, the form contained twenty-seven statements, which are listed in the Appendix. They were divided into four sections themed as: Behavior and Attitude; Emotions and Feelings; Learning Strategies; Interactions with the Teacher and Classmates. Students had to interpret the statements and mark one of the numeric options of the following Likert scale: 1 - Strongly disagree; 2 - Partially disagree; 3 - Neither disagree nor agree; 4 - Partially agree; 5 - Strongly agree.

An exploratory factor analysis was performed on the collected data in order to verify whether the four dimensions of engagement (behavioral, emotional, cognitive and social) were supported by the factor structure. We also evaluated the reliability of the total engagement scale and the individual scales for each identified factor. The results of the analysis, performed by the software SPSS8 (version 23) and JASP9 (version 0.13.1.0), are described in the following section.

The last section of the form was entitled “Open field for suggestions and comments” and requested students to help the teachers improve the ERT imposed by the COVID-19 pandemic. This was the only space for a discursive answer on the form.

Construction, presentation, analysis and interpretation of data

Out of the 172 students regularly enrolled in six classes, only 99 answered the questionnaire, which corresponds to about 58% of the students. In order to understand this rate of adherence to the evaluation invitation for the first module, it is important to consider that: 1st - the students had already done twelve ERT Physics activities; 2nd- both remote activities and research questionnaire were presented to students through forms. Thus, the rate of adherence to the questionnaire may have been limited by: (a) the tiredness of the students, who were demanded to answer successive forms; (b) the fact that the answer to the survey questionnaire is presented as anonymous and optional; (c) the fact that the students had already started studying a new group of subjects with which they were beginning to interact, for the first time, in the same period in which they were invited to answer the questionnaire.

Exploratory Factor Analysis

We decided to carry out an exploratory factor analysis of the data to assess whether the students gave coherent answers to the questionnaire items and whether the correlations between the responses supported a multidimensional conception of engagement. Our first step was to verify that our data met the necessary conditions to implement a factor analysis (Hair et al., 2009). The first condition concerns the sample size, which must be equal to or greater than 100. The 99 responses given to the questionnaire placed us at the lower limit of the first condition for this type of analysis.

The second condition is the MSA index, which varies from 0 to 1. The MSA must show values above 0.8. This index assesses the degree of intercorrelations between the variables and the adequacy of the sample. The MSA value for our data was 0.840. The third and last condition is that the correlations between the variables must have a significant value with p <0.05. The verification of this last condition was done with the Bartlett sphericity test, whose result was also shown to be adequate (X2 (351) = 1616.918; p <0.001).

Another important point of the exploratory factor analysis is the use of criteria to establish the number of factors or latent variables that adequately explain the correlations between the variables. Hair et al. (2009, p. 114) state that an exact quantitative basis for deciding the number of factors to be extracted has not been developed yet. For this reason, the decision on the number of factors maintained in a factor analysis must combine: (1) a conceptual foundation (how many factors should be in the structure?); (2) empirical evidence (how many factors can be reasonably sustained?). From this we conclude that the choice of the number of factors is related to the interpretation given to these factors, as well as their relevance in explaining the studied phenomena.

Not only from reading the bibliography on engagement, but also from following the process of developing the questionnaire, we expected to find four factors in the data analysis, but we knew that finding evidence to support them was necessary.

One of the criteria we used to find the number of factors was the parallel analysis. In the parallel analysis, several simulations are carried out with random data holding the same characteristics as the real data (same number of variables and subjects). A factor analysis is performed in each of these simulations and the mean eigenvalues of the factors obtained with the random data are compared with the eigenvalues of the factors of the analysis performed with the real data. The eigenvalues measure the percentage of variance explained by each factor, which must be greater than the one obtained just by chance. Therefore, it is considered that only factors with eigenvalues greater than those obtained from random data are significant.

Parallel analysis is usually more effective for identifying the number of factors than other methods (Hayton, Allen and Scarpello, 2004). The JASP program does this type of analysis and, therefore, we used it in our research. The result of this analysis showed that, in only three factors, the factorial analysis with real data had eigenvalues greater than those obtained from random data.

This result contradicted our initial expectation that the four dimensions of engagement mentioned in the work of Fredricks at al (2016) would emerge from an exploratory factor analysis. To interpret it, we have two hypotheses. The first is that the size of our sample was not large enough to allow the differentiation between the cognitive and behavioral dimensions of engagement that, in our data, appeared to be associated with a single factor, as we will see below. The second is that students may have interpreted the questionnaire items differently than we did. The first hypothesis is supported by the fact that the size of our sample was at the lower limit of what is necessary for a factor analysis. The second hypothesis cannot be properly appreciated, given that we did not conduct: (a) qualitative pilot studies whit focus groups; (b) interviews to assess how students interpreted the questions or to ask them to explain their answers.

At this point in the analysis, it is very important to clarify that the sample size may be sufficient to enable a factor analysis and, at the same time, not be large enough to allow a differentiation of all possible factors. This is a known fact in significance analysis: differences that are not significant in small samples can become significant if the sample size is increased.

Table 1 Factor loadings 

Factor 1 Factor 2 Factor 3 Uniqueness
Q_01 0,663 0,45
Q_02 0,534 0,553
Q_03 0,645 0,502
Q_04 0,757 0,362
Q_05 0,779 0,363
Q_06 0,751 0,405
Q_07 0,51 0,597
Q_08 -0,497 0,726
Q_09 0,623 0,564
Q_10 0,595 0,492
Q_11 0,58 0,542 0,281
Q_12 0,771 0,285
Q_13 -0,81 0,342
Q_14 -0,853 0,251
Q_15 -0,631 0,548
Q_16 0,758 0,374
Q_17 0,646 0,462
Q_18 0,654 0,446
Q_19 0,97
Q_20 -0,596 0,622
Q_21 0,505 0,598
Q_22 0,565 0,633
Q_23 0,524 0,56
Q_24 0,547 0,546
Q_25 0,651 0,497
Q_26 0,754 0,421
Q_27 0,499 0,7

Source: Developed by the authors using data from the JASP program (version 0.13.1.0).

Table 1 shows the result obtained by the JASP program for the factor loadings. The texts of statements Q_01 to Q_27 can be found in the Appendix. The exploratory factor analysis carried out by JASP identified three factors with the Varimax rotation method. Only loadings greater or equal to 0.40 were considered significant, according to the standards that are used in this type of analysis (Hair et al., 2009). Q_11 has loadings in two factors. Another dissonant highlight is statement Q_19, which has no factor loading. We decided to remove these two questions from our analysis.

In the Appendix, both are highlighted in gray lines, and they do not contain information in the cells presented in the third column. The discard of only two statements, coupled with the presence of twenty-five others that have a significant factor loading and are associated with only one of the three factors, evinces the existence of a consistent factor structure in the correlations. Together, the three factors identified in Table 1 explain 51% of the data variance. Although it seems low, this percentage is considered fully satisfactory for this type of analysis (Veiga, 2016; Gunuc and Kuzu, 2015; Nasir et al., 2020).

Our analysis allowed us to associate the factors identified in Table 1 with the dimensions of engagement proposed by Fredricks et al. (2016), although we found only three factors in our data. Factor 1 includes nine statements: Q_01, Q_07, Q_08, Q_10, Q_12, Q_13, Q_14, Q_15 and Q_20 (see Table 1). A careful reading of these statements, listed in the Appendix, shows that Factor 1 is related to the students’ feelings and emotions and, therefore, we associate it with Emotional Engagement.

A new query to Table 1 and to the Appendix shows that Factor 2 includes nine statements: Q_04, Q_05, Q_06, Q_09, Q_22, Q_23, Q_25, Q_26 and Q_27, and that they are all related to student interaction with the teacher or with their peers. For this reason, we associate Factor 2 with Social Engagement. Finally, Factor 3 contains seven statements: Q_02, Q_03, Q_16, Q_17, Q_18, Q_21 and Q_24, which are related to the students’ behaviors or their learning strategies. For this reason, we denominate Factor 3 as Cognitive-behavioral and interpret it as a combination of the behavioral and cognitive dimensions proposed in the literature on engagement.

The names of the three factors found in our factor analysis were introduced in the fourth column of the Appendix. It is this new denomination, and not the original classification of the statements, that must be taken into account in the continuation of this section dedicated to the presentation and analysis of the data.

2. Student engagement in all three dimensions

In the second stage of the statistical analysis, we added up the numerical values of the twenty five survey items provided by the students who remained after the discard of statements Q_11 and Q_19. To carry out the sum, we were careful to invert the values of the five survey items that corresponded to low engagement for values near 5. The items with this feature are shown in the penultimate column of the Appendix with an asterisk, in addition to being succeeded by the word Inverted between parenthesis. From this adjustment, the value obtained by the sum of the twenty five items validated by the factor analysis varied between 25 and 125. We used that value to establish a general student engagement scale.

After obtaining this general engagement scale, we started the third stage of the statistical analysis. This time, we gathered the seven survey items identified with the Cognitive-behavioral dimension and performed the same kind of sum as described above. We repeated this set of operations for the nine items associated with the Emotional dimension and for the nine items interpreted as an expression of the Social dimension of engagement.

In order to facilitate the comparison between the sums for the Cognitive-behavioral, Emotional and Social dimensions, we averaged the scales, that is, we divided the value of each scale by the number of items of each dimension in that scale. This way, we ensured that the value obtained for each factor or dimension lied within 1 and 5.

Table 2 presents the resulting general engagement scale mean and the three arithmetic means from the Cognitive-behavioral, Emotional and Social engagement scale dimensions.

Table 2 Statistics of engagement scales 

1- General 2- Cognitive-behavioral 3- Emotional 4- Social
Mean 3,42 3,75 3,45 3,16
Std. Deviation 0,76 0,83 0,94 1,06
Minimum 1,68 1,57 1,11 1
Maximum 4,88 5 4,89 5

Source: Developed by the authors using data from the SPSS program (version 23).

Besides the mean values, Table 2 presents the standard deviation and the maximum and minimum values of each measurement. Since numbers between 1 and 5 refer to a Likert scale, we decided to interpret the obtained values the following way: (a) values below 3 indicate low engagement, which is as lower as how close the value is to 1; (b) values above 3 indicate engagement, which as higher as how close the value is to 5.

To evaluate the reliability of the general engagement scale, as well as the engagement profile scales in each of the three dimensions shown in Table 2, we resorted to the calculation of Cronbach’s a coefficient (Hair et al., 2009). The result of this calculation, which measures the internal consistency of each scale, is shown in Table 3.

Table 3 Reliability of the scales 

Scale Number of items Cronbach’s coefficient
Total 25 0,92
Cognitive-behavioral 7 0,86
Emotional 9 0,89
Social 9 0,89

Source: Developed by the authors using data from the SPSS program (version 23).

Values for the Cronbach’s a coefficient above 0.8 indicate a consistent and highly reliable scale. Therefore, from Table 3 it is possible to see that the scales we have built are reliable and trustworthy.

3. Data and research questions

At this point of the analysis, we can go back to our research question. The data presented in the previous subsections are evidence of the reliability of the scales that we created to assess student engagement in the course, and also helped us qualify this engagement.

Considering the data in Table 2, we see that the mean values for Cognitive-behavioral, Emotional and Social engagement are different. Before interpreting these differences, we decided to check whether they were statistically significant. To achieve this, we performed the ANOVA test of repeated measures in SPSS, as well as a post hoc Bonferroni test, in the same program. Combined, these tests showed that all differences were significant [F(1.72; 168.63) = 17.57; p < 0.001]. Table 2 shows that Cognitive-behavioral engagement was the highest, followed by Emotional and Social.

The mean values of the answers given to each of the twenty-five statements collected by our instrument allow us to better understand how student engagement varied. If we consider that values greater or equal to 4.0 indicate high engagement, we see that there are: (a) three statements of high engagement (Q_17, Q_18 and Q_21) in the Cognitive-behavioral dimension (43% of the items);10 (b) two statements of high engagement (Q_08 and Q_20) in the Emotional dimension (22% of the items); (c) no statement of high engagement in the Social dimension.

The three statements of high engagement in the Cognitive-Behavioral dimension indicate that the majority of students: (a) tried to relate what they were studying with what they had learned from previous remote activities. (b) tried to understand their mistakes when they realized that they had made them; (c) strived to perform tasks and learn, even when they had difficulty. These statements are clear examples of what Fredricks et al. (2004) consider to be expressions of cognitive engagement and, more specifically, of metacognitive strategies to understand complex ideas. The two statements of high engagement in the Emotional dimension, in turn, indicate that most of the students: (a) were interested in the subjects; (b) did not give up on dealing with subjects that are difficult to understand or abandon them to perform other tasks.

We can interpret these findings as an indication that the activities of the Physics ERT allowed students to: (a) maintain their interest; (b) overcome the obstacles that they were faced with; (c) establish relationships between the studied phenomena and the concepts introduced to understand them.

A second way of interpreting the same findings, which does not contradict the previous one, concerns the students’ characteristics. In the section devoted to describe the research context of this manuscript, we drew attention to the fact that the students join the school through a selection process. This type of process tends to select students who have managed to remain engaged in their previous school experiences and who, for this reason, probably have the necessary conditions to learn Physics in the unfavorable conditions of the ERT provided by the school where we conducted the research.

Once we have addressed, in the previous paragraphs, our interpretation of the evidence of cognitive-behavioral and emotional engagement, we decided to end this section focusing on the evidence of low student engagement in the Social dimension. Only one statement associated with this dimension (Q_27) had a value significantly greater than three (average value of 3.84 in the Appendix). This is the statement in which students say they have used “other communication resources (WhatsApp, for example)” to interact with their peers. In a complementary way, statements Q_05 and Q_06, which have the two lowest averages among all items in the questionnaire, are precisely those that express teachers’ insistence on the use of chat rooms as an asynchronous interaction channel.

Data obtained in the discursive question of the questionnaire - which have not yet been mentioned in the analysis so far carried out - help us complement our interpretation of student engagement in the course. Thirty-three of the ninety-nine students who responded to the form used the “Open space for suggestions and comments”. We carefully read the responses whose content could be summarized in the following categories: positive and negative highlights of the course, suggestions, complaints and compliments. We shall focus on the most frequent instances.

Ten out of the thirty-three responses referred to the use of explanatory videos about the contents of the activities. The answers varied among: (a) compliments about the introduction of these resources from the middle of Module I on, as a result of a claim made by many students; (b) complaints about the absence of this resource at the beginning of the Module. Seven students used different words and language resources to ask teachers to “teach the subject” before asking questions. Ten other discursive responses presented suggestions or complaints on how teachers chose to interact with the classes: Google Chat. Six students addressed this issue by directly criticizing the mandatory use of the Chat. Four other students requested the use of WhatsApp instead of Chat.

In the presentation text of the second ERT Physics module, the teachers responded to some of the students’ requests in the open response questionnaire item. The teachers committed to adopt some of the suggestions and presented justifications for declining two specific requests. The first request refused was that to “explain the subjects before asking questions.”

As a justification for this refusal, teachers said it was important for students to understand how to answer questions about subjects that they do not fully comprehend yet because: (a) dealing with the discomfort provoked by this type of situation is an important part of the scientific learning process; (b) in order to study a science, it is necessary to learn how that science works and how scientists produce knowledge; (c) it is very common that, when studying science, there is a need to live with only partial answers to questions and to have to reason from questions, rather than from answers and reliable information. Although it is not our objective, in this paper, to analyze this pedagogical discourse of teachers, it is important to emphasize that it reveals important aspects of the conception of science teaching and learning in this group of teachers.

The second refusal from the teachers concerned the suggestion of using WhatsApp instead of Chat. The teachers stated that they had removed the mandatory use of Chat, but reiterated that the creation of a Chat room for each activity was the best way to allow students to count on their help in case of questions. To justify, they said that: (a) on Chat, it is possible to start a new conversation for each new question; (b) those who access the Chat and find conversations that have already been started can insert their question or comment in these conversations or start a new conversation with a different question; (c) in Chat it is easier to find answers to questions already addressed by the teacher; (d) none of these features exist on WhatsApp, where it is difficult to locate and resume an issue previously discussed.

Conclusion

The research hereby disclosed presented evidence of the engagement of students who attended an introductory, remote and emergency Physics course in adverse conditions. The results of the survey help to understand which aspects of the course elicited greater and lesser engagement.

Instead of the four dimensions of the engagement construct validated in the work of Fredricks et al. (2016), our exploratory factor analysis showed the behavioral and cognitive dimensions merged into a single factor.

This result can be interpreted not only as an effect of our sample size, but also as a result of the following limitations of our research: (a) we did not involve independent experts in the development of our construct;11 (b) we had a short time to perform the research and this made it impossible to carry out qualitative pilot studies with focus groups and/or interviews to assess how students interpreted the questions or to request explanations for the answers that these subjects gave to the instrument; (c) we did not triangulate our data with other sources of information (recordings of meetings held on Meet or analysis of interactions on Chats created for each activity, for example).

Being fully aware of those limitations, we consider that it is important to state that our research instrument: (a) was not submitted to a strong and complete validation; (b) cannot be used to measure student engagement in remote classes in other contexts. Even with those limitations, our research built evidence of student engagement and allowed us to qualify that engagement.

In addition, we would like to say that we work in cooperation with the team of teachers responsible for the course and that part of our data has had a significant and immediate impact on improvements in the following modules of the course. Accordingly, the main merit of our research is pedagogical and not theoretical or methodological. It is difficult for teachers to carry out research with short deadlines and in time for their results to interfere with planning and pedagogical practice. Hence the importance of collaboration between teams of teachers and researchers, whenever possible.

A featured characteristic of several engagement measurement or assessment instruments that we found was the adjustment of those instruments to the specificities of academic activities to which the students would or would not be engaged. This is a characteristic that hinders a complete and unadapt reuse of previously validated instruments in new contexts.

The evidence we found came from statistical analysis procedures accessed at two complementary programs: SPSS and JASP. The procedures used were presented and described at the fourth section of this paper. Our results indicate that a significant amount of students present engagement at essential aspects of learning associated with the Emotional and Cognitive-behavioral dimensions.

That is a remarkable result when considering the adverse conditions in which the ERT was instituted at the school.12 Regarding it, in the introduction of this paper we draw attention to the fact that several students accessed the proposed activities from their cellphones. On the other hand, in the section dedicated to present the theoretical framework and describe the methodology, we said that teachers had acknowledged their difficulty for turning a 100% face-to-face and strongly experimental teaching approach into a course mediated by asynchronous and remote activities.

As Dixson (2015), we believe that the engagement evaluation of distance or online courses can guide teachers to promote the required adaptations aiming to raise or increase the students’ involvement with their own educational experiences. Unlike this author, though, we believe that face-to-face courses have more and better resources than distance courses. Particularly, in our case, we consider that the context experienced by the students at the ERT were exceptionally unfavorable.

Thus, in spite of the enormous pedagogical effort undertaken by the teachers: (a) there was a sudden transformation from a 100% classroom-based and predominantly experimental course to a remote course; (b) laboratory classes with real equipment have become unviable, which has resulted in an impoverishment of the mediational resources used in the course; (c) the interpretation of phenomena and the use of ideas from the sciences could not continue to be carried out collaboratively through synchronous discursive interactions between the teacher and the students, and the students and their classmates; (d) there was a drastic reduction in the possibilities of students’ social interaction with their teachers and peers.

Thinking about the relevance of the work that we present here to the research community, we believe that one of its positive points was the consequent use of the possibilities offered by statistics for the understanding of educational phenomena, even in situations where it is not possible to carry out a methodologically robust research. Finally, we also believe that our research contributes to a new field of research dedicated to study the impacts of the COVID-19 pandemic on school education.

Considering the fact that we are still amidst a pandemic and the production and publication of research requires time help us understand the limitations of the quantity and quality of the works available on the topic when our paper was finalized. Some articles are experience reports (Barbosa, Ferreira and Kato, 2020; Hoffmann et al., 2020; Silva and Ramos, 2020; Piffero et al., 2020), others have methodological flaws that compromise the work (Médici et al., 2020). There are also those who present the imposition of remote education as an opportunity for “innovation towards education” without, however, displaying consistent data to corroborate this claim (Carneiro et al., 2020; Bezerra et al., 2020).

REFERENCES

Arruda, E. P. Educação Remota Emergencial: elementos para políticas públicas na educação brasileira em tempos de Covid-19. Em Rede, Revista de Educação a Distância, v. 7, n. 1, p. 257-275, 2020. [ Links ]

Bachelard, G. O novo espírito científico. Rio de Janeiro: Tempo Brasileiro, 2001. [ Links ]

Barbosa, R. e Jófili, Z. Aprendizagem Cooperativa e Ensino de Química - parceria que dá certo. Ciência e Educação, V. 10, No.1, p. 55-61, 2004. [ Links ]

Barbosa, A. T.; Ferreira, G. L.; Kato, D. S. O Ensino Remoto Emergencial de Ciências e Biologia em tempos de pandemia: com a palavra as professoras da Regional 4 da SBENBIO. REnBio - Revista de Ensino de Biologia da SBEnBio, V. 13, No. 2, p. 379-399, 2020. [ Links ]

Bezerra, K. P.; Costa, K. F. de L.; Oliveira, L. C.; Fernandes, A. C. L.; Carvalho, F. P. B.; Nelson, I. C. A. S. R. Ensino remoto em universidades públicas estaduais: o futuro que se faz presente. Research, Society and Development, v. 9, n. 9, 2020. [ Links ]

Chin, C. Teacher Questioning in Science Classrooms: Approaches that Stimulate Productive Thinking. Journal of Research in Science Teaching, Vol. 44, No. 6, PP. 815-843, 2007. [ Links ]

Carneiro, L. A.; Rodrigues, W.; França, G. Prata, D. N. Uso de tecnologias no ensino superior público brasileiro em tempos de pandemia COVID-19. Research, Society and Development , V. 9, No. 8, 2020. [ Links ]

Christenson, S. L.; Reschly, A. L. and Wylie, C. Handbook of Research on Student Engagement. Springer New York Dordrecht Heidelberg London, 2012. [ Links ]

Dixson, M. D. (2012). Creating effective student engagement in online courses: What do students find engaging?. Journal of the Scholarship of Teaching and Learning, 10(2), 1-13. [ Links ]

Dixson, M. D. Measuring Student Engagement in the Online Course: The Online Student Engagement Scale (OSE). Online Learning, [S.l.], v. 19, n. 4, july 2015. [ Links ]

Driver, R. et al. Construindo o Conhecimento Científico. Química Nova na Escola, n. 9, maio 1999. [ Links ]

Educação Vigiada 2020. Disponível em Disponível em https://educacaovigiada.org.br/ , acesso em 10/2020. [ Links ]

Fredricks, J.A., Blumenfeld, P.C., & Paris, A.H. School engagement: Potential of the concept, state of the evidence. Review of Educational Research, 74(1), 59-109, 2004. [ Links ]

Fredricks, J.A., Wang, M., Linn, J.S., Hofkens, T.L., Sung, H.C., Parr, A.K., & Allerton, J. Using qualitative methods to develop a survey measure of math and science engagement. Learning and Instruction, 43, 5-15, 2016. [ Links ]

Freitas. E. T. F; e Aguiar Júnior, O. G. A ação docente como sustentação da produção discursiva dos estudantes na sala de aula de Física de Educação de Jovens e Adultso. Revista Brasileira de Pesquisa em Educação em Ciências, Vol. 12, No1, 2012. [ Links ]

Gale M. Sinatra, Benjamin C. Heddy & Doug Lombardi (2015) The Challenges of Defining and Measuring Student Engagement in Science 2015, Educational Psychologist, 50:1, 1-13 [ Links ]

Gunuc, S., Kuzu, A. (2015). Student engagement scale: development, reliability and validity1. Assessment & Evaluation in Higher Education, 40, 587 - 610. [ Links ]

Hair Jr. , J. F, Rolph E. Anderson, Ronald. L Tatham e William C. Black. Análise Multivariada de Dados. 6. Ed. Porto Alegre: Bookman, 688 p., 2009. [ Links ]

Hayton, J. C., Allen, D. G., & Scarpello, V. (2004). Factor retention decisions in exploratory factor analysis: A tutorial on parallel analysis. Organizational Research Methods, 7, 191-205. [ Links ]

Hoffmann, W. P.; Loss, R. A.; Guedes, G. F.; Mexia, A. A.; Guedes, S. F. A importância do ensino remoto: Um relato da Universidade do Estado de Mato Grosso. Research, Society and Development , V. 9, No. 9, 2020. [ Links ]

Klem, A. M., & Connell, J. P. (2004). Relationships matter: Linking teacher support to student engagement and achievement. Journal of School Health, 74(7), 262-273. [ Links ]

Médici, M. S.; Tatto, E. R.; Leão, M. F. Percepções de estudantes do Ensino Médio das redes pública e privada sobre atividades remotas ofertadas em tempos de pandemia do coronavírus. Revista Thema, V.18, No Especial, p.136-155, 2020. [ Links ]

Nasir, M.A., Janikowski, T., Guyker, W.M., & Wang, C. (2020). Modifying the Student Course Engagement Questionnaire for Use with Online Courses. Journal of Educators Online, 17. [ Links ]

Paula, H. F.; Talim, S. L. Uso coordenado de ambientes virtuais e outros recursos mediacionais no ensino de circuitos elétricos. Caderno Brasileiro de Ensino de Física, v. 29, n. Especial 1, 2012. [ Links ]

Paula, H. F.; Talim, S. L. Avaliação de estudantes sobre a prática de produzir registros das atividades de ciências. Revista Ensaio, Belo Horizonte, v. 17, n. 1, p. 14-38, jan-abr, 2015. [ Links ]

Parra, H. Z. M; Cruz, L.; Amiel, T.; Machado, J. Infraestruturas, economia e política informacional: o caso do Google Suite for Education. Mediações-Revista de Ciências Sociais, v. 23, n. 1, p. 63-99, 2018. [ Links ]

Piffero, E. L. F.; Coelho, C. P.; Soares, R. G.; Roehrs, R. Metodologias ativas e o ensino remoto de biologia: uso de recursos online para aulas síncronas e assíncronas. Research, Society and Development , v. 9, n. 10, 2020 [ Links ]

Saraiva, K.; Traversini, C.; Lockmann, K. A educação em tempos de COVID-19: ensino remoto e exaustão docente Práxis Educativa, Ponta Grossa, V. 15, p. 1-24, 2020. [ Links ]

Silva, T. F.; Ramos, T. C. S. A graduação na área da saúde em tempos de pandemia da COVID-19: o ensino da disciplina de Prática de Promoção da Saúde e Prevenção de doenças. Research, Society and Development , V. 9, No. 9, 2020. [ Links ]

Tel Amiel, T.; Ewout ter Haar, E.; Vieira, M. S.; Soares T. C. Educação aberta, plataformas e capitalismo de vigilância: a pandemia como encruzilhada. In: LAVITS_COVID-19_#20, Setembro de 2020. Disponível em Disponível em https://lavits.org/lavits_covid19_20-educacao-aberta-plataformas-e-capitalismo-de-vigilancia-a-pandemia-como-encruzilhada/?lang=pt , acesso em11/2020. [ Links ]

Veiga, F. H. Assessing student Engagement in School: Development and validation of a four-dimensional scale. Procedia - Social and Behavioral Sciences, 217, 813-819, 2016. [ Links ]

Wang, M.-T., Fredricks, J. A., Ye, F., Hofkens, T. L., & Linn, J. S. The math and science engagement scales: scale development, validation, and psychometric properties. Learning and Instruction, 43, 16-26, 2016. [ Links ]

Wang, M.-T., Fredricks, J., Y e, F., Hofkens, T., & Linn, J. S. Conceptualization and assessment of adolescents’ engagement and disengagement in school: A Multidimensional School Engagement Scale. European Journal of Psychological Assessment, 35(4), 592-606, 2019. [ Links ]

Wang, Z., Bergin, C., & Bergin, D. Measuring engagement in fourth to twelfth grade classrooms: the Classroom Engagement Inventory. School psychology quarterly: the official journal of the Division of School Psychology, American Psychological Association, 29 4, 517-535. 2014 [ Links ]

2In the first year of the course, for example, students study alternating voltage circuits and notions of quantum physics.

3A significant part of the students did not have a computer, tablet or cell phone for exclusive use to access the ERE at predetermined times, as was the case in classroom teaching. For this reason, there was a need for the ERE to be predominantly asynchronous. This situation improved after the school helped students with difficulties accessing equipment or the internet.

4The school principals were probably unaware of the warnings offered by authors like Parra et al. (2018) about the risks of the phenomenon of over-concentration of information in digital technology giants. In addition, most teachers only learned of the school decision to join the Google Classroom too much later. Companies like Google have long invested “heavily” in developing applications for education. Travested as “free” software or maintained as proprietary software “borrowed” from schools and universities through agreements, the increasingly widespread use of these applications has been yet another source of profitable data for these companies. Authors such as Amiel, et al. (2020) and the Monitored Education report (2020) alert us that “free” software collects users’ personal data. The data is then sold to other companies and used to predict and induce behavior. This sustains the profits of the digital technology giants, while placing them as underpinnings of true surveillance capitalism.

5This intensified the transfer, to this private company, of personal data of teachers and students and data resulting from their academic activity. In addition, Google becomes the main form of access for students to a social right: education under the responsibility of a public school. This contradiction bothered several teachers at the school with whom the research team had contact through informal conversations. We did not refer to these conversations in the paper because they were not related to our research problem.

6The teacher’s master’s thesis is not included in the bibliographic references because such inclusion would imply the identification of the teacher and the school.

7It is worth remembering, in this respect, that this teacher also belongs to the team of teachers.

8SPSS is a package of statistical programs developed and marketed by the Statistical Product and Service Solutions Company.

9JASP is an open source statistical programs package developed and continuously updated by researchers at the University of Amsterdam.

10The statement Q_19, whose average also indicates high engagement, cannot be considered in this calculation, because it was not validated by factor analysis.

11Two PhD in Science Education, who were not part of the research group, participated in the development of the questionnaire. However, they cannot be considered independent experts because they are part of the team of teachers who used the questionnaire for educational purposes.

12The general conditions offered for teaching and learning in the public school where we carry out the research are quite favorable. These are better conditions than those found in most similar institutions and well above the reality of all public high schools. Most of the teachers in this school hold PhD degrees and work exclusively at there. In addition, most of them teach a maximum of twelve class hours per week. If these favorable and adequate working conditions do not guarantee the quality of education offered by the school, there is no doubt that they favor this quality.

1The vacancies in the selection process that allows students to join the school are divided into three parts. A part of the vacancies complies with the ethno-racial quota policy regulated by the Law. A second part is reserved for students from public schools and the third part is for students from private schools. The number of places reserved for these last two audiences is proportional to the number of entries in the selection. Historically, there are many more entries in the selection for students from public schools. As a result, there are more openings for these students.

APPENDIX: QUESTIONNAIREITEMS, ORIGINAL CLASSIFICATION, NEW CLASSIFICATION GENERATED BY FACTOR ANALYSIS AND MEAN RESPONSE VALUES

Classif. Orig. Classif. post analysis. Statement Mean
Q_01 Behavioral Emotional In most of the remote Physics activities from the module that just ended, I was able to stay focused on what I was doing. 3,28
Q_02 Behavioral Cognitive-behavioral In most of the remote Physics activities from the module that ended, I wrote down doubts and comments in order to consult them later. 3,11
Q_03 Behavioral Cognitive-behavioral In most of the remote Physics activities from the module that ended, I studied my notes and other materials to better understand the subject. 3,65
Q_04 Behavioral Social In most Chats created for remote Physics activities from the module that ended, I placed my questions. 3,14
Q_05 Behavioral Social In most of the Chats created for remote Physics activities from the module that ended, I commented on my colleagues’ questions. 2,51
Q_06 Behavioral Social In most of the Chats created for remote Physics activities from the module that ended, I made comments on the teacher’s answers. 2,22
Q_07 Emotional Emotional In most of the remote Physics activities from the module that ended, I was able to understand the importance of the subject for my life. 3,62
Q_08 Emotional Emotional *In most remote Physics activities from the module that ended, I had no interest in the subjects covered (Inverted). 4,01
Q_09 Emotional Social In most remote Physics activities from the module that ended, I enjoyed participating in Chats to interact with colleagues and the teacher. 3,36
Q_10 Emotional Emotional Most of the time, I kept a good expectation regarding the remote activities that would be proposed in the following weeks. 3,09
Q_11 Emotional ------------- In most of the remote Physics activities from the module that ended, I looked forward to the remote activities that would be proposed in the following weeks. 3,94
Q_12 Emotional Emotional In most remote Physics activities from the module that ended, I felt good while doing the activities. 3,06
Q_13 Emotional Emotional *In most remote Physics activities from the module that ended, I felt frustrated while doing the activities (Inverted). 3,26
Q_14 Emotional Emotional *In most remote Physics activities from the module that ended, I felt discouraged while performing the activities (Inverted.) 3,17
Q_15 Emotional Emotional *In most remote Physics activities from the module that ended, I felt bored while doing the activities (Inverted). 3,41
Q_16 Cognitive Cognitive-behavioral In most remote Physics activities from the module that ended, I studied the subjects covered to try to ensure a better understanding of the subject. 3,91
Q_17 Cognitive Cognitive-behavioral In most remote Physics activities from the module that ended, I tried to relate what I was studying to what I learned from previous remote activities. 4,17
Q_18 Cognitive Cognitive-behavioral In most remote Physics activities from the module that ended, I tried to understand my mistakes when I realized that I had given a wrong answer. 4,15
Q_19 Cognitive ------------- In most remote Physics activities from the module that ended, I preferred to receive the answers to the Physics questions from my colleagues or the teacher than to find them on my own (Inverted). 4,25
Q_20 Cognitive Emotional In most remote Physics activities from the module that ended, I gave up on dealing with difficult subjects to understand and I started to do the following tasks (Inverted). 4,12
Q_21 Cognitive Cognitive-behavioral In most remote Physics activities from the module that ended, I strived to perform the tasks and learn, even when I had difficulty. 4,38
Q_22 Cognitive Social In most remote Physics activities from the module that ended, I tried to work with my colleagues because I believed it would help me learn. 3,11
Q_23 Cognitive Social In most remote Physics activities from the module that ended, I used my colleagues’ ideas, questions or comments, which were posted on Chat, to better learn the subjects. 3,7
Q_24 Cognitive Cognitive-behavioral In most remote Physics activities from the module that ended, I tried to learn more and go beyond what was taught. 2,89
Q_25 Social Social In most remote Physics activities from the module that ended, I reached my teacher through Chat, E-mail or the Classroom Panel. 3,18
Q_26 Social Social In most remote Physics activities from the module that ended, I tried to help my colleagues who were having difficulties in learning. 3,35
Q_27 Social Social In most remote Physics activities from the module that ended, I used other communication resources (WhatsApp, for example) to interact with my colleagues, in addition to the Chat rooms created for each activity. 3,84

Received: December 04, 2020; Accepted: April 18, 2021

Contato: Helder de Figueiredo e Paula - Setor de Física do Colégio Técnico, Universidade Federal de Minas Gerais, Av. Antônio Carlos, 6627 - Pampulha, Belo Horizonte, MG | Brasil, CEP 31.270-901

Helder de Figueiredo e Paula - Phd in Education, Professor at “Colégio Técnico da UFMG”, Belo Horizonte, M.G., Brasil. E-mail: helder100@gmail.com.

Sérgio Luiz Talim - Phd in Physics, Retired Professor at “Colégio Técnico da UFMG”, Belo Horizonte, M.G., Brasil. E-mail: sergiotalim@gmail.com

Cecília Siman Salema - Undergraduate Physics College Student at ICEX/UFMG, Belo Horizonte, M.G., Brasil. E-mail: ceciliasiman15@gmail.com

Vinícius Reis Camillo - Undergraduate Physics College Student at ICEX/UFMG, Belo Horizonte, M.G., Brasil. E-mail: viniciusrcamillo@gmail.com.

Editor responsável: Glauco Santos Ferreira da Silva

Creative Commons License Este é um artigo publicado em acesso aberto sob uma licença Creative Commons