European Journal of STEM Education
Research Article
2021, 6(1), Article No: 05

Hybrid and Online Synchronous Delivery of Environmental Engineering During COVID-19 Pandemic: A Comparative Study on Perception, Attitude, and Assessment

Published in Volume 6 Issue 1: 07 Feb 2021
Download: 1592
View: 5125

Abstract

Hybrid and online synchronous delivery of courses is a time-demanding approach to teaching and learning systems that is designed to engage students in investigations of authentic concepts or problems without coming to the pre-assigned classrooms two or three times a week. This study presents perceptions and attitudes of students that attended a hybrid course in environmental engineering that suddenly converted to an online synchronous delivery due to COVID-19. It also presents an assessment of the hybrid and online synchronous delivery approach on the final exam scores as well as the final grades of the same course. The course, ‘Introduction to Environmental Engineering’, was developed as an online course for Civil and Environmental Engineering program students and delivered with hybrid and online synchronous options due to COVID-19 pandemic for several semesters to test the concept. In the hybrid delivery set up, all the quizzes and homework assignments were online while the midterm and final exams were in-class. For spring 2020 the final exam was online and for summer 2020 both the midterm and final exams were online due to COVID-19 adjustment. At the very end of the semesters, an online anonymous survey was conducted with five questions to understand the students’ perception and attitude on exam taking options and learning environment. Students’ perceptions and attitudes about online synchronous delivery approach compared to hybrid delivery approach, as well as the learning outcomes, appeared at face value not to be favorable. However, statistical analysis revealed that differences between online synchronous delivery and the Pre-COVID-19 hybrid delivery were not significant, indicating that, at least for these engineering students, online synchronous delivery is a justifiable option.

INTRODUCTION

Internet-based and online teaching and learning is becoming popular and was needed during COVID-19 pandemic. The relatively recent advent of Learning Management Systems (LMS), such as blackboard, Desire2Learn (D2L), eCollege, Moodle, and WebCT, and systems for offering lectures via MS Teams, Zoom and other platforms in the undergraduate setting in educational institutions has made it easy to provide online education, that is, web-based enhancement to traditional (face-to-face) classroom instruction (Rutter & Matthews, 2002). This online, hybrid or other mixed delivery approach allows instructors to combine the advantages of online class learning with the benefits of face-to-face interaction with relatively limited technological requirements on their part (Edling, 2000). The addition of a hybrid/online approach to the existing hybrid or face-to-face lecture-centric environmental engineering course is not supposed to reduce the quality of teaching and learning and would be welcomed and well received by students (Karim, 2016; Black, 2002). Preliminary reports suggest that the hybrid approach provides significant benefits for students and instructors, regardless of their level of technological expertise (Black, 2002; Brakels et al., 2002) and regardless of whether the classroom is hard-wired for live Internet access (Bento & Bento, 2000). Despite frequent use of an LMS for course administration purposes (content and lecture delivery), the instructors do not appear to be harnessing the full pedagogical potential of web-based augmentation via LMSs. The potential of LMS tools along with other online and mobile technology platforms to increase course administration/lecture delivery efficiency and maximize or enhance learning in traditional settings is an important educational issue that must be fully authenticated from both instructor and student perspectives (Papadakis et al., 2017; Papadakis et al., 2018; Woods et al., 2004). However, combining multiple modalities for online content delivery with a pot pouri of hybrid learning exercises that appeal to several learning styles may precipitate higher overall learning outcomes (McCray, 2000).

Theoretical Background

Traditional perception of teaching and learning in all levels is usually a face-to-face approach. Due to advancement of technology and internet connectivity teaching has evolved. In order to keep up with the technological advancement and new generations’ mindset traditional teaching moved away and web-based teaching found its way in the past decades. Although course delivery using either a hybrid or online approach may increase time demands and, in some cases, result in a loss of control, many instructors enjoy this approach because it allows for significant flexibility and benefits in instructions. Due to COVID-19 in March 2020 course delivery for many institutions worldwide had to change to online synchronous and/or asynchronous formats, and exams had to be administered online and/or use an alternative assessment process.

This study was designed to answer two questions: (1) What are the students’ perceptions and attitudes about the online synchronous course delivery along with online exam-taking options and online platforms as a learning environment? (2) Is there any effect of course delivery approach changes due to COVID-19 pandemic on students’ performance levels?

To answer the above questions, two objectives of this study were formulated. The first objective was to understand the students’ perceptions and attitudes about online synchronous course delivery along with online exam-taking options and online as a preferable learning environment for future environmental engineering courses. The second objective was to see the effect of COVID-19 change in the students’ performance levels. The first objective was accomplished via an anonymous online survey and the second objective was accomplished with statistical analysis of final exam scores, weighted average GPA, and the overall course grades. The goal of this study was to understand the overall effect of COVID-19 pandemic on students’ perceptions and attitude about an online synchronous course delivery and to assess the performance level, as well as to compare the performance levels between hybrid (Pre-COVID-19) and online synchronous (Post-COVID-19) course delivery options. The following sections outline the intervention, study methodology, instruments used, data collection and analysis, results and discussions, and finally conclusions and recommendations.

Intervention

Intervention simply means purposeful actions by a human agent to create and implement change. As we all know that end of 2019 and early in 2020, a pandemic of coronavirus (COVID-19) broke out in China and then spread globally. In the USA, spring semester starts in January and ends in May. Due to public health advisory and presidential Corona virus taskforce guidance, the education institutions in the USA had to make several changes in the course delivery in order to limit the spread of COVID-19. Within two weeks of this advisory, the educational institutions had to come up with an approach that would meet the guidance (6-ft social distance, washing hands, and face covering) without interruption of education. Several options were thought out such as converting all the courses to 100% online, either asynchronous or synchronous. Asynchronous delivery calls for video recording of lecture sessions and posting them in LMS. Since all faculty were not trained to be online instructors, synchronous option was chosen, with some training sessions for the faculty how to use MS Teams, Zoom, or Blackboard Collaborate for online synchronous delivery platform. That is how our university ended up delivering all courses online synchronous since mid-March 2020. The parameters for optimum group forming strategy, content design, effectiveness measurement, meta-analyses, impact of technology on PBL, implementation framework, alternative assessment in PBL, and other procedures for optimum learning that are acceptable to students and instructors were elaborated from the literature (Albanese & Mitchell, 1993; Berkson, 1993; Blumenfeld et al., 1991; Diehl et al., 1999; Frank et al., 2003; Karim, 2015; Krajcik et al., 1994; Moursund, 1999; Mills & Treagust, 2003; Ravitz & Blazevski, 2014; Strobel & Van Barneveld, 2009; Thomas et al., 1999; Vernon & Blake, 1993; Watson & Fang, 2012; Boubouka & Papanikolaou, 2013).

METHODOLOGY

The course, ‘Introduction to Environmental Engineering’, was developed as an online course for Civil and Environmental Engineering program students but taught as a hybrid (50% time in-class lecture and 50% time online self-paced) before COVID-19 (Pre-COVID-19) pandemic and online synchronous during COVID-19 (Post-COVID-19) pandemic for several semesters, to test the concept. In the hybrid course set up, all the quizzes and homework were online while the midterm and final exams were in-class. The spring 2020 final exam was online and during summer 2020 both midterm and final exams were online due to COVID-19 adjustment. The following two subsections outline the type of instruments used and the type of data collected and analyzed to accomplish the objectives of this study.

Instrument

To understand the effect of COVID-19 on the perceptions and attitudes of students learning (the first research question), an online anonymous survey via D2L was conducted at the end of the semester with five questions. The survey questions are presented in Figure 1. The first two questions were meant to investigate the students’ perceptions and attitudes about the course content and alignment delivered with online synchronous approach although no changes were made in the course content and other alignment. The third question probed an alternative assessment process using technologies and investigated students’ perceptions and attitudes about the challenges of conducting online closed book exams using a lockdown browser and webcam, as most of the traditional students were not familiar with these technologies. The fourth and fifth questions were to understand the students’ perceptions and attitudes about several test taking options (alternative assessments) and levels of learning. The second instrument used investigated the final exam scores, weighted average GPA, and the overall course grades to assess the students’ performance level and to compare the students’ learning environment between hybrid (Pre-COVID-19) and online synchronous (Post-COVOD-19) delivery (the second research question).

 

Figure 1. Survey questionnaire for hybrid/remote offerings of Environmental Engineering

 

Data Collection and Analysis

The data collected through the online survey was analyzed to understand students’ perceptions and attitudes about the course content and alignment, online exams using technologies, the exam-taking alternatives, and the degree of learning. The data that was collected and used to assess the performance levels was the final exam scores (maximum minimum, and average) and the overall course grades (Letter grades and weighted average GPA considering A = 4.0, B = 3.0, C = 2.0, and D = 1.0). F-grade was not included in the assessment as the students only receive an F-grade when they stop coming to the class or drop out after the deadline. The data was collected for the spring and summer 2020 semesters as online synchronous delivery and compared with data from the summer and fall 2019 semesters as hybrid delivery. There was a total of 34 students enrolled in summer 2019, 35 in fall 2019, 48 in spring 2020 (2 sections), and 33 in Summer 2020 semesters. Overall, 76 (about 51%) students participated in the survey for all 4 semesters. Nine students (about 26%) participated in the survey for summer 2019, 15 (about 43%) in fall 2019, 29 (about 60%) in spring 2020, and 23 (about 70%) in summer 2020. Seventy-four (74) students (about 49%) did not take the survey because the survey was not mandatory, and no incentive/grade points was given for taking the survey. Final exam scores, weighted average GPA, and the overall course grades were statistically analyzed and compared for differences between Pre-COVID-19 and Post-COVID-19 delivery. The analysis of data was performed with simple statistical tools and excel using goodness-of-fit tests such as ANOVA, χ2-tests, student t-tests, and F-tests, as necessary. The results of the data analysis are illustrated in the following section and in Figure 2 through Figure 6. Some of the responses to questions/options/choices, as seen in the figures, do not sum up to 100% as a few students did not respond to all questions or select all options or choices.

RESULTS AND DISCUSSIONS

Course Content and Alignment

For Q.1, overall, about 92% agreed that tests materials reflected what was covered in the class (Figure 2), both before and during the COVID-19 situation. The highest score was for the summer 2020 class (96%), followed by spring 2020 (93%), summer 2019 (89%), and fall 2019 (87%).

 

Figure 2. Distributions of responses for Q.1

 

The distributions of Q.2 responses are presented in Figure 3. It can be seen that, overall, about 93% of the students, participating in the survey, agreed that is there is a good agreement between the course outline and the course content, with the highest score (100%) for summer 2019, followed by fall 2019 and spring 2020 (93%), and summer 2020 (91%). It is apparent that students’ perceptions and attitudes about the course content and the alignment were consistent and similar for the Pre-COVID-19 and Post-COVID-19 situation.

 

Figure 3. Distributions of responses for Q.2

 

Online Exam Using Technologies

The weighted average response to Q.3, as to how the participants liked to take online midterm and final exams through D2L using Respondus Lockdown Browser and Webcam, was 2.83, which is close to the middle, with only a minority (25%) of students being positive and declining from spring to summer 2020. (see Figure 4; there were of course no values for summer 2019 and fall 2019). About 7% of the participants did not answer this question. It is clear that online examination with lockdown browser and webcam is not popular, presumably because it can be cumbersome to make the lockdown browser and webcam work, depending on the computer and the individual knowledge of computer operations.

 

Figure 4. Distributions of choices of the participants for Q.3

 

Alternative Evaluation and Learning

Regarding Q.4 on preferences for the midterm and final exams, 17% of the participants chose option 1: “Get the questions from D2L, print it, take it, scan and submit it in submission folder without proctoring”, 37% chose option 2: “Take-home exam for a day or two”, 13% chose option 3: “100% online with Multiple Choice Questions like a quiz”, and 25% chose option 4: “100% online that is get the questions in D2L like a quiz, take the exam like quiz and do the detail work in papers proctored using webcam, scan the papers in pdf and submit the papers in submission folder” (see Figure 5). About 8% participants did not answer this question. Option 2 (take-home exam) has the highest score, whereas students’ preference for online quizzes seem to decline from spring to summer.

 

Figure 5. Distributions of responses of the participants for exam delivery options (Q.4)

 

This could indicate that the students prefer to complete the test in their own time with the possibility of external help, as it has to be open book; and that this creates less anxieties than being watched by someone or a webcam. However, this cannot be confirmed until a take-home exam is conducted and evaluated.

To check the face-value outcomes above, a chi-square goodness-of-fit test was performed to validate or reject the null hypothesis “no differences from semester to semester and among four exam-taking options”. The chi-square test data are shown in Table 1. From the chi-square test, a p-value of 0.2913 was obtained, which is greater than both 0.05 (α = 5%) and 0.01 (α = 1%). A χ2-value of 3.7371 was also obtained. For a degree of freedom of 3, the critical values for χ2 are 7.81 (for α = 5%) and 11.3 (for α = 1%). The chi-square (χ2) value is less than the critical values of both the significance levels. So, the null hypothesis cannot be rejected, and it cannot be concluded that the differences from semester to semester and among the four exam taking options are statistically meaningful.

 

Table 1. Chi-square goodness-of-fit test for Q.4 data

Semester

Observed Values

Expected Values

Option 1

Option 2

Option 3

Option 4

Total

Option 1

Option 2

Option 3

Option 4

Total

Spring 2020

4

10

6

9

29

5.438

11.479

4.229

7.854

29

Summer 2020

5

9

1

4

19

3.563

7.521

2.771

5.146

19

Total

9

19

7

13

48

9

19

7

13

48

p-value = 0.2913; χ2-value = 3.7371; DF = 3, χ2-critical = 7.81 (for α = 5%) and 11.3 (for α = 1%)

 

To verify this, a single factor ANOVA was performed, and the data is presented in Table 2. Since F < Fcritical (in this case, 2.1538 < 6.5913), the null hypothesis indeed cannot be rejected.

 

Table 2. ANOVA for Q.4 data

Group

Sum

Count

Average

Variance

Source

SS

DF

MS

F

p-value

F-crit

Option 1

9

2

4.5

0.5

Between group

42

3

14

2.154

0.2362

6.591

Option 2

19

2

9.5

0.5

Within group

26

4

6.5

---

---

---

Option 3

7

2

3.5

12.5

Total

68

7

---

---

---

---

Option 4

13

2

6.5

12.5

 

 

 

 

 

 

 

 

Regarding Q.5 on learning, 29% of the participants chose option 1: “Learned same as hybrid/face-to-face”, 15% chose option 2: “Learned more online than hybrid/face-to-face”, and 48% chose option 3: “Learned less online than hybrid/face-to-face”. About 8% participants did not answer this question. The distributions of Q.5 responses are presented in Figure 6 and it seems obvious that students think they learned less online than in the hybrid or face-to-face conditions.

 

Figure 6. Distributions of responses of the participants for learning materials options (Q.5)

 

To see the variations of the three learning options for spring 2020 and summer 2020 (summer 2019 and fall 2019 were not included as these semesters were not a part of the remote offerings), a chi-square goodness-of-fit test was performed to validate or reject the null hypothesis “no differences from semester to semester and among three learning options”. The chi-square test data are shown in Table 3. The p-value is too high and the χ2-value is too low to reject the null hypothesis, so the differences between the semesters and among the three learning options are not statistically significant.

 

Table 3. Chi-square goodness-of-fit test for Q.5 data

Semester

Observed Values

Expected Values

Option 1

Option 2

Option 3

Total

Option 1

Option 2

Option 3

Total

Spring 2020

11

5

13

29

9.063

4.833

15.104

29

Summer 2020

4

3

12

19

5.938

3.167

9.896

19

Total

15

8

25

48

15

8

25

48

p-value = 0.4063; χ2-value = 1.8015; DF = 2, χ2-critical = 5.99 (for α = 5%) and 9.21(for α = 1%)

 

To verify this, a single factor ANOVA was performed, and the data are presented in Table 4. Since F < Fcritical (in this is the case, 4.0555 < 9.5521), it was confirmed that the null hypothesis cannot be rejected.

 

Table 4. ANOVA for Q.5 data

Group

Sum

Count

Average

Variance

Source

SS

DF

MS

F

p-value

F-crit

Option 1

15

2

7.5

24.5

Between group

73

2

36.5

4.055

0.1403

9.552

Option 2

8

2

4.0

2.0

Within group

27

3

9.0

---

---

---

Option 3

25

2

12.5

0.5

Total

100

5

---

---

---

---

 

Assessment

The analysis of assessment outcomes was based on the final grades for hybrid (summer 2019 and fall 2019) and online (spring 2020 and summer 2020) exam options. The data are presented in Table 5. From the chi-square test, a p-value of 0.0401 was obtained, which is less than 0.05 (α = 5%) but greater than 0.01 (α = 1%). A χ2-value of 17.5994 was also obtained. For a degree of freedom (DF) of 9, the critical values for χ2 are 16.9 (for α = 5%) and 21.7 (for α = 1%). Therefore, with α set at 1%, the null hypothesis cannot be rejected but when α is set at 5%, it can. So, with some uncertainty, it can be concluded that the differences in the final grades from semester to semester and between hybrid and online exam options are significant.

 

Table 5. Assessment based on final grades using Chi-square Goodness-of-fit test

Exam Option

Semester

Observed Grades

Expected Grades

A

B

C

D

Total

A

B

C

D

Total

Hybrid
(Pre-COVID-19)

Summer 2019

10

14

7

3

34

8.04

14.01

9.65

2.30

34

Fall 2019

9

18

5

1

33

7.80

13.60

9.36

2.23

33

Online
(Post-COVID-19)

Spring 2020

9

13

22

4

48

11.35

19.78

13.62

3.24

48

Summer 2020

7

16

8

2

33

7.80

13.60

9.36

2.23

33

 

Total

35

61

42

10

148

35

61

42

10

148

p-value = 0.0401; χ2-value = 17.5994; DF = 9, χ2-critical = 16.9 (for α = 5%) and 21.7 (for α = 1%)

 

A second analysis of assessment was based on the weighted average GPA for hybrid and online exam semesters. The data is presented in Table 6. From the chi-square test, a p-value of 0.9974 was obtained which is greater than both 0.05 (α = 5%) and 0.01 (α = 1%). A χ2-value of 0.0459 was also obtained. For a degree of freedom of 3, the critical values for χ2 are 11.1 (for α = 5%) and 15.1 (for α = 1%). The chi-square (χ2) value is less than the critical values of both 7.81 (α = 5%) and 11.3 (α = 1%). Therefore, the null hypothesis cannot be rejected so the differences in GPA from semester to semester and between hybrid and online exam options are not statistically significant. A t-Test and an F-Test performed for this parameter confirmed this.

 

Table 6. Assessment based on weighted average GPA using Chi-square Goodness-of-fit test

Exam Option

Semester

Observed GPAs

Expected GPAs

 

Hybrid (Pre-COVID-19)

Summer 2019

2.9118

2.8458

p-value = 0.9974

 

c2 value = 0.0459

Fall 2019

3.0606

2.8458

Online (Post-COVID-19)

Spring 2020

2.5625

2.8458

Summer 2020

2.8485

2.8458

 

Total

11.3834

11.3834

 

The third analysis of assessment was based on the final exam Minimum, Average, and Maximum scores obtained by students in the hybrid and online delivery semesters. The analysis is presented in Table 7. From the chi-square test, a p-value of 0.9959 was obtained which is greater than both 0.05 (α = 5%) and 0.01 (α = 1%). A χ2-value of 1.3995 was also obtained. The critical values for χ2 are 12.6 (for α = 5%) and 16.8 (for α = 1%) for a degree of freedom of 6. The chi-square (χ2) value is less than the critical values of both significance levels. Therefore, the null hypothesis cannot be rejected, and it cannot be concluded that differences between exam scores from the semester to semester and between hybrid and online exam options are significant.

 

Table 7. Assessment based on the final exam scores using Chi-square Goodness-of-fit test

Exam Option

Semester

Observed Values

Expected Values

Min

Avg

Max

Total

Min

Avg

Max

Total

Hybrid
(Pre-COVID-19)

Summer 2019

30

58

88

176

30.05

56.98

88.96

176

Fall 2019

30

67

100

197

33.64

63.78

99.58

197

Online
(Post-COVID-19)

Spring 2020

35

55

95

185

31.59

59.90

93.51

185

Summer 2020

30

57

87

174

29.71

56.34

87.95

174

 

Total

125

237

370

732

125

237

370

732

p-value = 0.9659; χ2-value = 1.3995; DF = 6, χ2-critical = 12.6 (for α = 5%) and 16.8 (for α = 1%)

 

The summary of the goodness-of-fit test analyses is listed in Table 8 for four different types of data. Based on the goodness-of-fit test and ANOVA it was apparent that students’ perceptions and attitudes about the four different exam-taking options and three different learning options did not differ significantly, although face-value analysis of the responses suggested otherwise. The analyses of the final exam scores (maximum, minimum, and average), weighted average GPA, and the overall final grades confirmed that the observed differences were not big enough to say that outcomes differed significantly. The only value that proved significant (and only at the 5% level) was the difference between the final grades for the hybrid (2019) and the online (2020) exams.

 

Table 8. Summary of Goodness-of-fit test analyses

Data Type: c2-Test

p-value

c2-value

DF

Critical Value

c2-Test Comment

0.05

0.01

Students’ choices for four exam options (Survey – Q.4)

0.2913

3.7371

3

7.81

11.3

p-values are greater than both 0.05 (a = 5%) and 0.01 (a = 1%) and c2-values are less than the corresponding critical values. The differences are not significant.

Students’ choices for three learning options (Survey – Q.5)

0.4063

1.8015

2

5.99

9.21

Weighted average GPA

0.9974

0.0459

6

12.6

16.8

p-values are greater than both 0.05 (a = 5%) and 0.01 (a = 1%) and c2-values are less than the corresponding critical values. The differences are not significant.

Final exam scores (Min., Avg, Max.)

0.9659

1.3995

6

12.6

16.8

Final Grades

0.0401

17.5994

9

16.9

21.7

p-value is less than both 0.05 (a = 5%) and greater than 0.01 (a = 1%) and c2-value is greater than 16.9 (a = 5%) and p-value is greater than 0.01 (a = 1%) and c2-value is less than 21.7 (a = 1%). Differences are significant at the 5% uncertainty level.

Data Type: t-Test

p-value

t-value

DF

tcritical
(two tail)

t-Test Comment

Weighted average GPA

0.2237

1.7413

2

4.3026

Since t-value is within -tcritical and +tcritical, the null hypothesis cannot be rejected. The observed values are not big enough to conclude that differences are significant.

Data Type: F-Test

p-value

F-value

DF

Fcritical
(one tail)

F-Test Comment

Final Grades

0.3055

0.2708

1

0.0069

Since F-value > Fcritical, the null hypothesis is rejected. Therefore, variances of the two populations, hybrid and online exam options, are NOT equal.

Data Type: ANOVA

p-value

F-value

DF

Fcritical

ANOVA Comment

Students’ choices for four exam options (Survey – Q.4)

0.2362

2.1538

3

6.5914

Since F < Fcritical (in this is the case, 2.1538 < 6.5914), therefore, the null hypothesis cannot be rejected. The four exam taking options are statistically equal.

Students’ choices for three learning options (Survey – Q.5)

0.1403

4.0555

2

9.5521

Since F < Fcritical (in this is the case, 4.0555 < 9.5521), therefore, the null hypothesis cannot be rejected. The three learning options are statistically equal.

 

CONCLUSIONS AND RECOMMENDATIONS

From the data of this study, it can be concluded that a project-based or problem-based learning (PBL) course with alternative assessment options without physical contact, such as oral assessment or take-home exam can be introduced and implemented during COVID-19 without harm to the students’ learning and performance. In this study, an effort was made to understand the students’ perceptions and attitudes with regard to the learning environment as well as their performance levels in environmental engineering for the changes in the course offerings due to COVID-19 in the middle of spring 2020. Students agreed that the course content in the online format was aligned with the content of assessment, but their perceptions and attitudes about learning in online environment and taking online exams using technologies (respondus lockdown browser and webcam) appeared to be not favorable. Students’ choice was the take-home exam. Fortunately, it could not be proved by statistical analysis that the online synchronous approach significantly degraded the level of students’ performance, although face value analysis suggested that online synchronous delivery approach does not maintain the same level of students’ performance. It is the author’s opinion that PBL delivery with take-home exam, as preferred by the students in this study, along with other alternative evaluation processes (cf. Boubouka & Papanikolaou, 2013) can be adopted to maintain the students’ learning and performance at pre-COVID-19 levels.

STUDY LIMITATIONS

A source of bias for this study could be the fact that the author was the only person who designed this study, conducted the survey, collected the semester end data, analyzed the data and had, as a teacher, an interest in a positive outcome. Another limitation is that all students were in the same engineering program, making it unclear whether the outcomes would be the same for other STEM students. The last limitation is the number of students involved to test the hypotheses. Perhaps, with a larger sample, face value differences do become significant.

Figure 1 Figure 1. Survey questionnaire for hybrid/remote offerings of Environmental Engineering
Figure 2 Figure 2. Distributions of responses for Q.1
Figure 3 Figure 3. Distributions of responses for Q.2
Figure 4 Figure 4. Distributions of choices of the participants for Q.3
Figure 5 Figure 5. Distributions of responses of the participants for exam delivery options (Q.4)
Figure 6 Figure 6. Distributions of responses of the participants for learning materials options (Q.5)
  • Albanese, M. A., & Mitchell, S. (1993). Problem-based learning: A review of literature on its outcomes and implementation issues. Academic Medicine-Philadelphia, 68(1), 52–81.
  • Bento, R. F., & Bento, A. M. (2000). Using the Web to extend and support classroom learning. College Student Journal, 34(4).
  • Berkson, L. (1993). Problem-based learning: Have the expectations been met? Academic Medicine, 68(10), S79-S88. https://doi.org/10.1097/00001888-199310000-00053
  • Black, G. (2002). A comparison of traditional, online, and hybrid methods of course delivery. Journal of Business Administration Online, 1(1), 1–9.
  • Blumenfeld, P. C., Soloway, E., Marx, R. W., Krajcik, J. S., Guzdial, M., & Palincsar, A. (1991). Motivating project-based learning: Sustaining the doing, supporting the learning. Educational Psychologist, 26(3–4), 369–398. https://doi.org/10.1207/s15326985ep2603&4_8
  • Boubouka, M., & Papanikolaou, K. A. (2013). Alternative assessment methods in technology enhanced project-based learning. International Journal of Learning Technology, 8(3). https://doi.org/10.1504/IJLT.2013.057063
  • Brakels, J., van Daalen, E., Dik, W., Dopper, S., Lohman, F., van Peppen, A., Peerdeman, S., Peet, D. J., Sjoer, E., & van Valkenburg, W. (2002). Implementing ICT in education faculty-wide. European Journal of Engineering Education, 27(1), 63–76. https://doi.org/10.1080/03043790110100191
  • Diehl, W., Grobe, T., Lopez, H., & Cabral, C. (1999). Project-based learning: A strategy for teaching and learning. Boston, MA: Center for Youth Development and Education, Corporation for Business, Work, and Learning.
  • Edling, R. J. (2000). Information technology in the classroom: Experiences and recommendations. Campus-Wide Information Systems, 17(1), 10-15. https://doi.org/10.1108/10650740010317014
  • Frank, M., Lavy, I., & Elata, D. (2003). Implementing the project-based learning approach in an academic engineering course. International Journal of Technology and Design Education, 13(3), 273–288. https://doi.org/10.1023/A:1026192113732
  • Karim, M. A. (2015). Project Based Learning of Environmental Engineering: A Case Study. Proceedings of ASEE’s 122nd Annual Conference & Exposition. https://doi.org/10.18260/p.24607
  • Karim, M. A. (2016). Hybrid Delivery of Environmental Engineering: Perception, Attitude, and Assessment. Proceedings of the 2016 ASEE 123rd Annual Conference and Exposition. https://doi.org/10.18260/p.25506
  • Krajcik, J. S., Blumenfeld, P. C., Marx, R. W., & Soloway, E. (1994). A collaborative model for helping middle grade science teachers learn project-based instruction. The Elementary School Journal, 94(5), 483–497. https://doi.org/10.1086/461779
  • McCray, G. E. (2000). The hybrid course: Merging online instruction and the traditional classroom. Information Technology and Management, 1(4), 307–327. https://doi.org/10.1023/A:1019189412115
  • Mills, J. E., & Treagust, D. F. (2003). Engineering education—Is problem-based or project-based learning the answer. Australasian Journal of Engineering Education, 3(2), 2–16.
  • Moursund, D. G. (1999). Project-based learning using information technology. International society for technology in education Eugene, OR.
  • Papadakis, S., Kalogiannakis, M., Sifaki, E., & Vidakis, N. (2017). Access moodle using smart mobile phones. A case study in a Greek University. In Interactivity, Game Creation, Design, Learning, and Innovation (pp. 376–385). Springer. https://doi.org/10.1007/978-3-319-76908-0_36
  • Papadakis, S., Kalogiannakis, M., Sifaki, E., & Vidakis, N. (2018). Evaluating Moodle use via Smart Mobile Phones. A case study in a Greek University. EAI Endorsed Transactions on Creative Technologies, 5, 156382. https://doi.org/10.4108/eai.10-4-2018.156382
  • Ravitz, J., & Blazevski, J. (2014). Assessing the role of online technologies in project-based learning. Interdisciplinary Journal of Problem-Based Learning, 8(1), 9. https://doi.org/10.7771/1541-5015.1410
  • Rutter, L., & Matthews, M. (2002). InfoSkills: A holistic approach to online user education. The Electronic Library, 20(1), 29-34. https://doi.org/10.1108/02640470210418245
  • Strobel, J., & Van Barneveld, A. (2009). When is PBL more effective? A meta-synthesis of meta-analyses comparing PBL to conventional classrooms. Interdisciplinary Journal of Problem-Based Learning, 3(1), 44–58. https://doi.org/10.7771/1541-5015.1046
  • Thomas, J. W., Mergendoller, J. R., & Michaelson, A. (1999). Project-based learning: A handbook for middle and high school teachers. Novato, CA: The Buck Institute for Education.
  • Vernon, D. T., & Blake, R. L. (1993). Does problem-based learning work? A meta-analysis of evaluative research. Academic Medicine, 68(7), 550-563. https://doi.org/10.1097/00001888-199307000-00015
  • Watson, W. R., & Fang, J. (2012). PBL as a framework for implementing video games in the classroom. International Journal of Game-Based Learning (IJGBL), 2(1), 77–89. https://doi.org/10.4018/ijgbl.2012010105
  • Woods, R., Baker, J. D., & Hopper, D. (2004). Hybrid structures: Faculty use and perception of web-based courseware as a supplement to face-to-face instruction. The Internet and Higher Education, 7(4), 281–297. https://doi.org/10.1016/j.iheduc.2004.09.002
AMA 10th edition
In-text citation: (1), (2), (3), etc.
Reference: Karim MA. Hybrid and Online Synchronous Delivery of Environmental Engineering During COVID-19 Pandemic: A Comparative Study on Perception, Attitude, and Assessment. European Journal of STEM Education. 2021;6(1), 05. https://doi.org/10.20897/ejsteme/9688
APA 6th edition
In-text citation: (Karim, 2021)
Reference: Karim, M. A. (2021). Hybrid and Online Synchronous Delivery of Environmental Engineering During COVID-19 Pandemic: A Comparative Study on Perception, Attitude, and Assessment. European Journal of STEM Education, 6(1), 05. https://doi.org/10.20897/ejsteme/9688
Chicago
In-text citation: (Karim, 2021)
Reference: Karim, M. A.. "Hybrid and Online Synchronous Delivery of Environmental Engineering During COVID-19 Pandemic: A Comparative Study on Perception, Attitude, and Assessment". European Journal of STEM Education 2021 6 no. 1 (2021): 05. https://doi.org/10.20897/ejsteme/9688
Harvard
In-text citation: (Karim, 2021)
Reference: Karim, M. A. (2021). Hybrid and Online Synchronous Delivery of Environmental Engineering During COVID-19 Pandemic: A Comparative Study on Perception, Attitude, and Assessment. European Journal of STEM Education, 6(1), 05. https://doi.org/10.20897/ejsteme/9688
MLA
In-text citation: (Karim, 2021)
Reference: Karim, M. A. "Hybrid and Online Synchronous Delivery of Environmental Engineering During COVID-19 Pandemic: A Comparative Study on Perception, Attitude, and Assessment". European Journal of STEM Education, vol. 6, no. 1, 2021, 05. https://doi.org/10.20897/ejsteme/9688
Vancouver
In-text citation: (1), (2), (3), etc.
Reference: Karim MA. Hybrid and Online Synchronous Delivery of Environmental Engineering During COVID-19 Pandemic: A Comparative Study on Perception, Attitude, and Assessment. European Journal of STEM Education. 2021;6(1):05. https://doi.org/10.20897/ejsteme/9688
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Submit My Manuscript



Phone: +31 (0)70 2190600 | E-Mail: info@lectitojournals.com

Address: Cultura Building (3rd Floor) Wassenaarseweg 20 2596CH The Hague THE NETHERLANDS

Disclaimer

This site is protected by copyright law. This site is destined for the personal or internal use of our clients and business associates, whereby it is not permitted to copy the site in any other way than by downloading it and looking at it on a single computer, and/or by printing a single hard-copy. Without previous written permission from Lectito BV, this site may not be copied, passed on, or made available on a network in any other manner.

Content Alert

Copyright © 2015-2024 LEUKOS BV All rights reserved.