Remote elementary education (Part 2): grade-level comparative analysis with control

Purpose – The purpose of this two-part research was to investigate the effect of remote learning on student progress in elementary education. Part 2, presented in this paper, is a follow-up study to examine how student progression in the two pandemic-induced environments compared to the pre-pandemic conditions. Design/methodology/approach – The authors expanded the quantitative, quasi-experimental factorial design of the authors ’ initial study with additional ex-post-facto standardized test score data from before the pandemic to enhance the group comparison with a control: the conventional pre-pandemic classroom environment. Thus, the authors were able to examine in which ways the two pandemic-induced learning environments (remote and hybrid) may have affected learner progress in the two subject areas: English Language(ELA)andMath.Additionally,theauthorsprovidedagrade-by-gradebreakdownofanalysisresults. Findings – Findings revealed significant group differences in grade levels at or below 6th grade. In the majority of analyzedcomparisons, learner achievement in the hybrid group was significantly lowerthan those in either the remote or the classroom group, or both. Research limitations/implications – The additional findings further supported the authors ’ initial hypotheses: Differences in the consistency and continuity of educational approaches, as well as potential differences in learner predispositions and the availability of home support systems may have influenced observed results. Thus, this research also contributes to the general knowledge about learner needs in elementary education. Originality/value – During the pandemic, remote learning became ubiquitous. However, in contrast to e-learninginpostsecondaryeducation,forwhichanabundanceofresearchhasbeenconducted,relativelylittleisknownabouttheefficacyofsuchapproachesinelementaryeducation.

In previous research (Kleinke and Cross, 2021), we investigated the effect of remote learning on student progression in two subject areas crucial to elementary education: English Language (ELA) and Math. Brought on by the circumstances of the global pandemic, many schools across various nations were forced to suspend in-person, in-classroom learning and adopt some form of remote education (Storey and Slavin, 2020). Once gradually returning to in-person learning, educators were faced with difficult decisions about how to best balance the constantly changing public health guidelines, parent expectations and student needs, which often resulted in schools offering learning simultaneously in different modalities (Kleinke and Cross, 2021).

Remote elementary education 2
259 Therefore, our previous quasi-experimental research (Kleinke and Cross, 2021) focused on the analysis of group differences in standardized test scores between fully remote and hybrid learners. We found a significant difference in learning progression between the two groups for both of the investigated subject areas. Furthermore, the findings suggested that fully remote students outperformed their hybrid-learning peers, which we tentatively attributed to the greater consistency in the fully remote environment. While students in the fully-remote environment continued their learning through pandemic lock-down and gradual re-opening unchanged, students in the hybrid environment experienced frequent changes to learning conditions and school policies, which may have negatively affected their learning achievements (Kleinke and Cross, 2021).
Thus, the question that arose from our initial research (Kleinke and Cross, 2021) was whether the observed group differences were indeed caused by negative effects in the hybrid environment and not attributable to performance increases in the fully remote environment. Accordingly, the purpose of this follow-on study was to further examine how student progression in the two pandemic-induced environments compared to the pre-pandemic conditions. Therefore, we expanded the quantitative, quasi-experimental factorial design of our initial study with additional ex-post-facto test score data from before the pandemic to create a third condition for comparison: the conventional classroom environment. Based on our tentative theory from our initial findings (Kleinke and Cross, 2021), we hypothesized: H1 0 . There are no significant differences in student development among the three different learning environments.
H1 a. Student development in the hybrid learning environment (H) is significantly less than under both fully remote (R) and conventional classroom (C) conditions. Furthermore, the revised research design also allowed us to investigate whether any observed group differences would be consistent across all grade levels, which would suggest that learners are equally affected by the conditions regardless of age. We, therefore, also hypothesized that: H2 0 . The distribution of any group differences is consistent across all grade levels.
H2 a . Group differences vary across grade levels with younger learners more affected by learning conditions.

Literature review
Consistency and continuity in teaching and learning have long been identified as important to student success. Founded on Dewey's (1938) principle of continuity of experiences, learning environments, pedagogical approaches and academic curricular traditionally sought to provide a consistent learning experience in order to maximize student achievement (Cantley et al., 2021). While his philosophy also emphasized the importance of personal experiences, Dewey (1938) stressed that individual interactions are not enough if not carefully choreographed into a consistent learning experience by a skilled educator. Nevertheless, research conducted at the beginning of the new millennium (e.g. Driessen and Sleegers, 2000;Jamieson and Wikeley, 2000) started to challenge this conventional view of the role of consistency as essential to effective learning, maintaining that adaptation to individual needs may be more crucial. For example, Driessen and Sleegers (2000), in their study of 7,410 8th grade students from 567 schools in the Netherlands found that consistency of the teaching approach was of no relevance to the measured levels of achievement, and socio-ethnic background and the prevalence of disadvantaged students were much more important predictors for performance.
At the same time, the home learning environment (HLE) has also been increasingly identified by researchers as an important factor for elementary children's academic and social development (Lehrl et al., 2020a, b). For example, through multilevel stepwise regression analysis, Dimosthenous et al. (2020) found that, in addition to teacher factors, certain HLE aspects such as home-learning activities and home-learning materials had a significant shortterm, as well as some long-term, effect on student achievements. They also concluded that some of the potential deficits in a student's HLE may be offset by teacher factors (i.e. that a good teacher can make up for some of the negative effects of a less than ideal learning environment at home; Dimosthenous et al., 2020). Furthermore, these studies seem to point to the need for increased collaboration between a child's home and their school to positively affect learning outcomes (Lehrl et al., 2020a, b), seemingly supporting both concepts: individualized attention as well as consistent delivery.
When applied to remote learning, educators must adopt effective methods of instruction and reach consistency in their teachings. Consistency may help teachers as well as students to feel more organized and less stressed (O'Leary et al., 2021). Consistency may also help students to feel more relaxed and comfortable in the classroom, which might increase their participation and engagement. The more consistent a teacher can be, the more trust and respect students and their parents may have for that teacher and expectations can become mutual. Thus, consistency seems especially important during the early years of education when learning effectiveness depends on the above-discussed home-to-school connections (Lehrl et al., 2020a, b). Van Ameijde et al. (2018) reported that consistency of teaching methods and styles during remote learning seems to result in more effective mastery of math and language skills in the majority of students in primary school; therefore, particularly in an elementary school remote learning environment, such as during the pandemic, teachers need to adopt effective approaches that help them to reach consistency (Gunzenhauser et al., 2021).
To reach consistency in remote learning, it is important to curate resources in one centralized location or database. The more time students and parents spend looking for assignments and resources, the less likely it is that learning will occur. In a sense, in a remote learning environment, the previously mentioned HLE aspects of home-learning activities and home-learning materials (Dimosthenous et al., 2020) seem to become a shared responsibility between the remote facilitators and the students' home support system. In this regard, school leaders also play an important role in the process, as they have to encourage their teachers to curate the content they already use in a way that is easy for parents and students to reference (Murai and Muramatsu, 2020). Singh et al. (2021) investigated the preparedness of faculty with little or no experience teaching online courses during the COVID-19 pandemic. Using a fishbone analysis, the authors were able to identify root problems which were then used to create a strength, weaknesses, opportunities and challenges (SWOT) matrix for their online teaching experiences. Singh et al. (2021) concluded that as online and hybrid teaching will increase in the future, not only is better teacher preparation needed, but support from other stakeholders, including administrators, parents and students, seems crucial to provide meaningful learning experiences. Together, all stakeholders should aim to reach consistency and continuity in remote elementary learning.
Especially in times of a pandemic, when remote learning becomes a good, if not essential, alternative to the conventional classroom setting (Ferri et al., 2020;Storey and Slavin, 2020), one size may not fit all when it comes to instruction, and teachers should have the flexibility to make in-the-moment decisions to personalize learning. However, as Schafer et al. (2021) point out, in times like these, it is often more important to be consistent than it is to be creative. Teachers involved in remote learning should establish daily parameters for learning. Both students and parents are still adapting to working and learning from home. To aid in that transition, educators can set daily schedules and clear learning objectives to easily identify what students should be focusing on each day. As educators navigate these changes in Remote elementary education 2 instructional practices, it is critically important for both administrators and teachers to present consistent information. If leaders do not have a clear vision for instruction, teachers may be left to their own devices, which could cause students and parents to try keeping track of multiple different platforms for daily learning. Dhawan (2020) researched educational institutions in India and their response to the shift away from traditional teaching methods during the COVID-19 pandemic. Conducting a SWOT analysis, Dhawan (2020) concluded that a strong IT infrastructure and e-learning platform are critical to successful implementation of online learning. Once the infrastructure and platform are established, people seem more accepting of the online teaching environment.
Daily teaching routines should also look consistent, especially for students at the primary grade levels who may not be using digital resources on a daily basis (Schafer et al., 2021). Students report that communication and materials delivery become an issue during a remote study if teachers change their teaching styles. Thus, consistency in teaching methods further seems important because it helps to compensate for a lack of personal connection. Especially during a pandemic, each community may face different challenges, making it difficult to remain consistent across the board (Fontenelle-Tereshchuk, 2021). However, consistency may help to ensure equity and access for all students regardless of grade level or socioeconomic status. The application of remote learning strategies offers opportunities for students and educators to learn without the direct, face-to-face interaction (Armfield et al., 2021). Nevertheless, in such a situation, educators need to have effective strategies, and providing consistency and continuity in the learning environment seems to be one important aspect.

Research method
For this follow-on research, the initial group comparison (Kleinke and Cross, 2021) between elementary students that spent the entire pandemic year under remote learning conditions (remote learning group, coded as R for the grouping variable) and those who returned to some form of in-person learning, albeit with restrictions such as mask mandates, at some point during the assessment cycle (hybrid learning group, coded H) was expanded by introducing a third learning condition: the conventional pre-pandemic classroom (classroom learning group, coded as C for the grouping variable and serving as the control). To measure student achievements in the three groups, pre-existing assessment data from the three consecutive years 2019, 2020, and 2021 were utilized, which allowed establishing within-subject paired pretest-posttest samples for each group. Thereby, pre-test scores (at the beginning of each assessment period) served as covariate moderating any observed differences in the posttest assessments (at the end of each assessment period). The 2019 and 2020 test results (both assessments from before the pandemic) served respectively as pretest and posttest variables for the C group, while 2020 and 2021 test results (one assessment from just before pandemic lockdowns and one from after gradual reopening) provided pretest and posttest scores for the two treatment groups R and H. In this way, the assessment data were collected and tracked for both subject areas Math and ELA.
By further splitting the data into grade levels, a comparison of R and H to a control C became possible: For example, the pre-pandemic 2019/2020 assessment data for all students that were in the 2020/2021 academic year in Grade 2 became the conventional classroom control C against which the 2020/2021 Grade 1 students in the R and H groups were compared. Thus, the analysis design accounted for grade-level-specific achievement advances in an academic year. Similarly, the within-subject paired pretest-posttest design in each group took individual pre-treatment differences among students and across conditions into consideration, allowing to assess relative learner development rather than just their absolute achievements. As previously discussed in our initial research (Kleinke and Cross, 2021), such design was important for ruling out any influences of (self-) selection biases since group assignments were essentially non-random.
As for the first part of our study (Kleinke and Cross, 2021), the setting for this research was a medium-sized charter school in north-central Colorado, which offers K through 12 education and regularly conducts standardized assessments of student performance utilizing Curriculum Associates' (2021) i-Ready test. This i-Ready assessment is a grade-level-based, norm-referenced, adaptive test for students' ELA and Math abilities that is widely used in elementary schools across the US (Bunch, 2017;Curriculum Associates, 2021;Ezzelle, 2017). Its recurrent, computer-based application was beneficial to the study since it ensured consistent testing conditions across all three learning environments a well as regular assessment data (three time per academic year: September, January and May) over multiple years. Another advantage of using i-Ready as the research instrument was that the January 2020 test administration happened just immediately before schools were forced to suspend inperson learning, making it a valuable pre-test tool for this research (Kleinke and Cross, 2021).
The sample for this study consisted of the assessment data for N 5 904 students from grade levels K through 8. While our initial study (Kleinke and Cross, 2021) focused on analyzing group differences between R and H students across the entire data set, for this follow-on research we further broke down the assessment data into grade levels, which also allowed us to add pre-pandemic conventional classroom comparison data. Nevertheless, due to the varying availability of these assessment scores in the different grade levels, individual samples sizes varied widely during the analysis. For example, while the C group for 3rd grade Math scores contained N C,3,Math 5 81 paired pretest-posttest samples, the 8th grade R group in ELA had only N R,8,ELA 5 6 samples available. Thus, differences in group sizes as well as relatively small sample sizes for 7th and 8th grade were some of the limitations to the study that potentially affected the confidence in results in some of the groups and comparisons.
As outlined in our initial research methodology (Kleinke and Cross, 2021), all statistical analysis was performed using the jamovi software tools (The jamovi project, 2021), including applicable plug-in packages and visualizations (Ben-Shachar et al., 2020;Fox and Weisberg, 2020;Gallucci, 2019;Lenth, 2020;R core Team, 2021;Singmann, 2018) and the preparation of the data included thorough examinations of descriptive statistics to identify outliers and test for analyses assumptions. A particular emphasis during the preparation and treatment of the data was put on identifying and eliminating any potential effects of differences in testing. An analyses of covariance (ANCOVA) on the posttest i-Ready scores (2021 scores for groups R and H and 2020 scores for group C) was conducted for each grade level (K through 8), utilizing the three-level (R, H, or C) grouping variable as factor and the paired pretest scores (Jan 2020 scores for groups R and H and Jan 2019 scores for the C group) as the covariate. For those grade levels for which pre-test data were not available (see discussion of limitations), the treatment was reduced to an analysis of variance (ANOVA) among the three groups. Similarly, for the ELA subject area, for which control group pretest data were not available (see discussion of limitations), both an ANCOVA between only the two groups H and R as well as a posttest-only ANOVA between all three groups was conducted for each grade level. In addition to the analyses for each grade level and subject, Math scores across all grade levels combined were also analyzed for their development over time in the H and R groups via a repeated measure ANOVA of 2019, 2020 and 2021 scores. This time series, while less specific than individual grade level analyses, served to gain an overall picture of how students' achievements in the different learning environments during the pandemic compared to, and progressed from, the year preceding it.

Assumptions, limitations and delimitations
As previously highlighted, one limitation of the methodology we used to follow up our initial analysis (Kleinke and Cross, 2021) was that the grade-level breakdown resulted in uneven Remote elementary education 2 group sizes as well as small samples sizes in the upper grade levels. Another limitation to our design was that no 2019 ELA data existed since the school did not use i-Ready for assessments of ELA development during that year. Therefore, ELA classroom control groups (C) across all grade levels lacked pretest covariate data, making the assessment of student development in Math the more robust analysis of the two. Furthermore, obviously, there were also no pretest data available for the K grade level since these students just started their education during the pandemic, and no ELA testing was conducted at the K grade level during any year. Therefore, analysis of K grade level data was mainly a measure of achievement rather than student development and was limited to Math only. Additionally, this lack of K-grade ELA data also, obviously, affected availability of 1st-grade pretest data in ELA, highlighting again that ELA was the less robust of the two measures in our analysis (Kleinke and Cross, 2021).
As also already delimitated in our initial study (Kleinke and Cross, 2021), the purpose of this research was not to identify any individual differences in students' achievement scores or determine their potential causes (e.g. cheating). Rather, the selected research design and analysis methods were intended to eliminate or control, as much as possible, for any differences in groups not associated with the different learning environments, essentially aiming to isolate and quantify the true effect that the learning environment may have had on student development. In this way, generalizable conclusions can be drawn that allow for better administrative decision-making under difficult circumstances such as the pandemic. It should also be clearly delimitated again that this research solely concentrated on the effects the learning environment may have had on academic achievements in the selected two subject areas. Other effects such as on social-emotional well-being or peer interactions were not within the scope of this research but may be valid considerations when making educational decisions that have holistic student development in mind. Furthermore, findings of this research may be specific to the particular setting, in which the school district and individual administrators were able to provide an adequate level of technical support to all learner groups. Therefore, the study did not take into account any potential issues or differences in accessibility or support for the remote learners, and results may be vastly different in environments in which such disparities exist (Kleinke and Cross, 2021).

Results
As previously outlined in the methodology discussion, to follow up our initial investigation of differences between R and H groups in the combined data (Kleinke and Cross, 2021), i-Ready score relationships were further examined at individual grade levels (K through 8), including a third comparison group C as the prior year control where available (see discussion of limitations). This detailed analysis by grade level and subject area indicated: There was a significant difference between groups for the posttest-only ANOVA in K Though, as mentioned, R group size was relatively small and no pretest data were available to confirm group equivalency or account for pretest differences via covariate. As discussed, there were also not sufficient ELA data available at the K grade level to conduct any analysis for that subject area (see Figure 1). For the ANCOVA testing of 1st grade Math, the 2020 Math Pretest scores were a significant predictor of 2021 Math Posttest scores, F(1, 122) 5 216.48, p < 0.001, r 5 0.79, and grouping was, as well, a significant factor, F(2, 122) 5 3.68, p 5 0.028. Nevertheless, there was also a significant interaction effect between group and pretest, indicating inhomogeneous regression slopes (Field, 2009). Once accounted for, pairwise post-hoc testing revealed no significant difference between the R (N R,1,Math 5 26, M R 5 413, SE 5 2.76) and C (N C,1,Math 5 62, M C 5 415, SE 5 1.73) groups (t R-C [120] 5 À0.649, p 5 n.s.); however, the H group (N H,1,Math 5 38, M H 5 408, SE 5 2.22) scored significantly lower than the C group (t H-C [120] 5 À2.498, p 5 0.014). The difference between R and H groups was insignificant (t H-R [120] 5 À1.39, p 5 n.s.). Overall, the grouping factor had a medium effect size (partial η 2 5 0.122; see Figure 2).  As previously discussed, testing for ELA in 1st grade was again posttest ANOVA only and revealed a significant difference in scores between groups, F(2, 104) 5 10.8, p < 0.001, η 2 5 0.172. Pairwise post-hoc comparison indicated that the H group (N H,1,ELA 5 57, M H 5 441, SE 5 6.53) scored significantly lower (t H-R [104] 5 À4.649, p < 0.001) than the R group (N R,1,ELA 5 40, M R 5 488, SE 5 7.79). Differences between the C group (N C,1,ELA 5 10, M C 5 415, SE 5 1.73) and both other groups were insignificant (t H-C [104] 5 À0.976, p 5 n.s.; t R-C [104] 5 1.766, p 5 n.s.); though, the sample size for C was relatively small, leading to a larger uncertainty in the confidence interval (see Figure 3).
For the posttest-only ANOVA of 2nd grade ELA Posttest (2021) scores, grouping was not significant, F(2, 182) 5 2.24, p 5 n.s., η 2 5 0.024. However, pretest group equivalency could not be verified since insufficient ELA Pretest data existed, not just for the C group but in all groups. Therefore, there was also no ANCOVA performed for 2nd grade ELA.
While, the lower grade analyses up to this point all had some drawbacks and inconsistencies and, consequently, only allowed limited conclusions, the following results for the middle grade levels in the data set were the most complete and, therefore, most fundamental to arriving at generalizable conclusions that may explain the overall observed effect. In general, assessment data for 3rd through 6th grade were the most comprehensive, providing adequate sample sizes and best satisfying statistical assumptions. Results, here,   Similarly, ANCOVA testing of 3rd grade ELA Posttest scores between H and R groups indicated that the 2020 ELA Pretest scores were also a significant predictor, F(1, 84) 5 69.13, p < 0.001, r 5 0.65, and grouping was a significant factor, F(1, 84) 5 8.30, p 5 0.005. Post-hoc testing indicated again that the H group (N H,3,ELA 5 58, M H 5 534, SE 5 4.25) scored significantly lower (t H-R [84] 5 À2.88, p 5 0.005) than the R group (N R,3,ELA 5 29, M R 5 555, SE 5 6.02).
ANCOVA testing of 4th grade ELA Posttest scores between H and R groups indicated that the 2020 ELA Pretest score was, as well, a significant predictor, F(1, 83) 5 102.39, p < 0.001, r 5 0.74, but grouping was marginally insignificant, F(1, 83) 5 3.15, p 5 n.s. Further examining the ELA posttest scores for 4th grade in all three groups (H, R, and C), the posttest- the grouping factor had a medium effect size (partial η 2 5 0.109; see Figure 10).
In contrast, ANCOVA testing of 5th grade ELA Posttest scores between H and R groups indicated that the 2020 ELA Pretest scores were, as well, a significant predictor for 2021 performance, F(1, 70) 5 126.19, p < 0.001, r 5 0.79, but grouping was marginally insignificant, F(1, 70) 5 3.90, p 5 n.s. Nevertheless, there was also a marginally significant interaction effect between group and pretest, that, once accounted for, resulted in a significant difference (t H-R [69] 5 À2.25, p 5 0.028) between H group (N H,5,ELA 5 57, M H 5 602, SE 5 3.03) and R group (N R,5,ELA 5 16, M H 5 617, SE 5 5.78) scores (see Figure 11). Further examining the ELA posttest scores for 5th grade with the posttest-only ANOVA across all three groups, there were no significant differences between groups, F(2, 217) 5 1.99, p 5 n.s. Nevertheless, also noteworthy was that sample sizes vastly differed between ANCOVA and ANOVA testing since there existed far fewer 2020 ELA Pretest scores than 2021 ELA Posttest ones, leading to a much smaller number of complete score pairs for ANCOVA testing than data points for the posttest-only ANOVA.
For 6th grade Math, the ANCOVA testing indicated that the 2020 Math Pretest scores were also a significant and strong predictor of 2021 Math performance, F(1, 112) 5 268.232, p < 0.001, r 5 0.84, but grouping was not a significant factor, F(2, 112) 5 0.48, p 5 n.s. Similarly, the ANCOVA testing of 6th grade ELA data revealed no significant differences between H and R groups, F(1, 96) 5 0.327, p 5 n.s., only confirming again 2020 ELA Pretest scores as a significant predictor, F(1, 96) 5 190.41, p < 0.001, r 5 0.81, and the posttest-only ANOVA testing across all three groups was also insignificant, F(2, 200) 5 2.27, p 5 n.s. This trend of no significant differences between groups continued into the upper grade levels; though, sample sizes also started to dramatically decrease here.
Finally, a repeated measures ANOVA was conducted on the entire data set, comparing the year-over-year development in student i-Ready Math scores between the H and the R groups for the time series 2019, 2020, and 2021. Post Hoc testing revealed a marginally significant difference (t[406] 5 À1.99, p 5 0.047) for the grouping factor for the 2021 results, while 2019 and 2020 did not differ significantly. The score development over time for both groups is depicted in Figure 12.

Discussion
Although, the statistical results of the time series ANOVA were not the strongest of all of the analyses that were conducted (likely because the aggregated data of all the grades contained also upper grade level data with little differences as well as year-over-year test score development differences among grade levels that contributed to increased uncertainty intervals), their graphical depiction may paint the clearest picture of what likely happened: While the score development of fully remote students essentially followed a continuous line through the three years, hybrid students' score development decreased in slope between 2020 and 2021 (i.e. during the pandemic), indicating that the pre-pandemic development of yearover-year learning performance could not be maintained for that group.
Similarly, analysis results for grades 3 through 6 (for which, as previously mentioned, the data were the most complete) consistently indicated that the hybrid group (H) performed significantly lower than the remote group (R), the prior-year classroom control group (C), or both. Furthermore, this observed trend can also be found in most of the lower grade levels below third grade, though, as previously discussed, the conclusiveness of these results was more limited due to the unavailability of either covariate or control group comparison data, or both. The only exception to this trend was in second grade Math, for which the prior-year control group (C) scores were somewhat lower than in both other groups (R and H). This specific case seemed to be an outlier in the study, which potentially could have been caused by Group T1 H R Figure 12. Estimated marginal means of math test score development over time in the two different groups from the repeated measures ANOVA below average performance of second grade students on the standardized test in 2020 (i.e. low 2020 test results would negatively affect C scores but also inflate H and R ones due to their use as a covariate for those two groups).
Therefore, in general, the grade-wise analysis seems to indicate that the hybrid learning environment (with all its pandemic-induced inconsistencies) negatively affected student achievements in both Math and ELA for students at or below 6th grade. This observed effect seemed to diminish in grades 7 and 8, leading to the tentative conclusion that the learning environment has the greatest effect on student achievements in elementary education, though generalized conclusions about the higher grade levels in the study are more limited due to the limited availability of data in those grades. Therefore, both H1 0 and H2 0 were rejected for this study.
Furthermore, R performance seemed, overall, consistent with the achievements of the prior-year control group (C) in almost all analyses for which control group data were available. Thus, there was no indication that the fully-remote environment significantly differed from the conventional classroom environment concerning the learner development in the two subject areas Math and ELA. The only exception to this observed trend was in the K-grade Math data, for which scores in the R significantly exceeded those of both other groups (H and C; though, as previously mentioned, group size and unavailability of pre-test covariate data make this result less reliable). Such anomaly could have been caused by an exceptional teacher or the additional support that K-grade remote students received at home, or both. In either case, nowhere in this study did students in the R group perform significantly worse than the classroom control group (C), suggesting that a comparative level of subjectspecific education can be achieved in fully-remote elementary education, even under adverse conditions such as during a global pandemic.
Therefore, what remains to be discussed is our perspective on the implications of the observed trends. As previously highlighted, through the research design and the analysis methods, we tried to completely and consistently eliminate any effects that the testing (e.g. abnormalities during test-taking at home) or non-random group assignments (i.e. the selfselection of the remote learning group) may have had on analysis results. By eliminating outliers and employing a control group design, as well as pre-test covariate data, we are confident that the observed group differences are, indeed, a genuine effect of the learning environment on student development. Furthermore, we observed that the biggest differences in those two learning environments during the pandemic (H and R) were likely caused by two main groups of factors: Environmental factors, which included aspects such as comfort, familiarity/experience, attention/care/support and consistency, all of which may have contributed to students and their parents electing to continue fully remote even though a hybrid option was available. These fully remote students likely stayed fully remote because their HLE could support this modality, providing a consistent learning experience that properly coordinated responsibilities between the remote facilitators and the students' home support system. Furthermore, the school that served as the setting for this study afforded a consistent approach to remote learning, fully embracing the modality and supporting the remote teachers and learners, as well as providing the necessary support structures, equipment, and processes.
In contrast, at least some of the students returning to hybrid learning may have had less choice due to the lack of a sufficient HLE and support for remote learning. They also may have faced a higher degree of inconsistencies due to ever-changing public health guidelines for the return to in-person learning. Thus, our findings seem to be consistent with recent research that stresses the importance of a supportive HLE (e.g. Dimosthenous et al., 2020;Lehrl et al., 2020a, b) and, in contrast to Driessen and Sleegers (2000) or Jamieson and Wikeley (2000), we found evidence for the benefits of consistency and continuity in the learning Remote elementary education 2 environment. Additional research could focus on student and parent perceptions about the available learning choices, as well as environmental factors such as the availability of home support, that may have contributed to the decision to either continue remotely or return to hybrid learning. Furthermore, follow-on research should examine whether and when the observed group differences equalize again, as learning returns to normal, allowing to also investigates any long-term effects the pandemic may have had on learners.
Developmental factors, which included aspects such as the development of responsibility, discipline, independence and individual decision-making/choice all of which may have contributed to students remaining fully remote if their learning preferences were supported in such an environment. While not the immediate focus of this study and its literature review, the ability to self-direct and self-regulate one's learning has been previously identified as crucial to online education in older learners (Association for the Study of Higher Education [ASHE], 2014). Thus, it seems plausible that similar prerequisites and learner dispositions may also hold true for elementary students in a remote learning environment. Therefore, if students chose to remain fully remote because this modality best aligned with their learning preferences, and such a fitting learning environment allowed the students to excel despite of the pandemic circumstances, it would also support the importance of individualized learning approaches that cater to student needs and predispositions. Though, further research, especially into the role of learning preferences in younger learners, seems necessary.

Conclusion
Our findings from this second part of the analysis seem to confirm our tentative conclusions presented in the first part (Kleinke and Cross, 2021), namely that environmental and developmental factors played a major role in student performance in the different learning environments during the pandemic. Consistency and continuity in the fully remote environment seemed to allow the learners to achieve a performance comparable to the prepandemic classroom control. In contrast, inconsistencies in the hybrid environment, as well as reduced home support, may have contributed to lower achievement levels for learners in that environment. Furthermore, the observed effects seemed to decrease with learner age and showed no significant differences in higher grade levels. More research is needed to confirm contributing factors to the observed differences between learning environments, as well as any long-term consequences of the environment choice.