Student Choice, Instructor Flexibility:
Moving Beyond the Blended Instructional Model
Jackie B. Miller
The Ohio State University
Mark D. Risser
The Ohio State University
Robert P. Griffiths
The Ohio State University
As postsecondary enrollments in the United States continue to increase—2000 to 2010 saw a 41% increase in undergraduate enrollments (Knapp, Kelly-Reid, & Ginder, 2012)—there is pressure for traditional brick-and-mortar universities to provide more diverse delivery options for students to take courses. Given the popularity of strictly online education, one choice for these institutions is to use hybrid learning methods, in which courses consist of synchronous online and face-to-face components. One promising hybrid learning model is the HyFlex (hybrid, flexible) model, originally designed by Dr. Brian Beatty (2010) for his graduate courses at San Francisco State University.
The HyFlex Model
The primary feature of the HyFlex model is to combine synchronous online and face-to-face components ("hybrid") in a single course and allow students to choose when and how they attend ("flexible"). Beatty (2010) defines HyFlex courses to be those that "enable a flexible participation policy for students whereby students may choose to attend face-to-face synchronous class sessions or complete course learning activities online without physically attending class" (Introduction, para. 1).
In a HyFlex course, the choice of participation mode can be made independently by each student. The instructor provides lecture content, structure, and activities to meet the goals of the syllabus. This must be done in such a way as to give "equivalency," so that students can experience the course content and complete the course requirements in comparable ways whether attending online, face-to-face, or in some combination of the two (Beatty, 2010). A true HyFlex course should also maximize accessibility by ensuring that students are equipped with the technological skills needed to access the participation choices (Beatty, 2010).
Though it has existed for nearly a decade (Beatty, 2006), relatively little research exists about the effectiveness of the HyFlex model for student learning within higher education. Thus far, much of the information around HyFlex stems from Beatty’s work, which has been featured at a variety of conferences, including the 2010 Educause Learning Initiative, and was identified as an Effective Practice in 2008 by the Sloan Consortium (2012). The formal applications of HyFlex are mostly concentrated in the California State University system, where the model is useful due to the fact that many students work full time or commute long distances to attend class (Beatty, 2010). There is little research into the application of HyFlex to large courses, where it can arguably provide significant cost-savings and other benefits. The question of how to provide an engagingly interactive experience in a course with several hundred students remains unanswered—particularly when a portion of those enrolled are not sitting in the physical lecture space.
Comparisons between HyFlex and related models
A variety of other pedagogical models have experimented with ideas similar to those used in the HyFlex model. Live lecture streaming (LLS), which involves the "broadcasting of a lecture over the internet [sic] at the same time as it is being delivered in the traditional lecture theater" (Fernando, Cole, Tan, & Freitas, 2011, p. 2), is a course model similarly motivated by cost, demand, and cohort size issues. Lecture recordings are made available for students who are unable to attend lecture during its scheduled time, as well as to enable remediation (Fernando et al., 2011). The more generically-titled "video streaming" model implemented by Buhagiar and Potter (2010) delivers a fully online, synchronous video lecture to the enrolled students, but also gives students the option of attending the live (traditional) version of the class that is being broadcasted (Buhahiar & Potter, 2010).
These models have many of the same motivations and components as the HyFlex model, but many studies in this area focus on distance learning, not hybrid methods for local students (Abdous & Yoshimura, 2010). HyFlex also enrolls all students in a single course section, instead of separating sections by delivery mechanism, as both LLS and the video streaming model do. HyFlex allows flexibility in attendance method on a daily basis, accommodating a broad variety of learning preferences and, more practically, the diversity of college students’ busy schedules. Furthermore, all students in a HyFlex course have equal opportunity to interact with the instructor and other classmates; the extent and nature (i.e., in person or online) of this interaction can also vary based on individual preference and need. Therefore, while containing some overlap, the HyFlex model is sufficiently distinct as to warrant independent treatment.
Backchannel Communication and Audience Response Systems
Educators can choose from numerous technologies when implementing the HyFlex model. These choices will have a major impact on the suitability of the model for a given situation, and on the ultimate success of the implementation. This study supplemented basic video streaming and recording tools with two additional technologies intended to amplify the viability and equivalency of available attendance options: backchannel communication and an audience response system (ARS).
Backchannel communication. "Backchannel" is a mode of communication created by audience members to connect with other observers both inside and outside of the presentation space, with or without the speaker’s awareness (Atkinson, 2009). The backchannel—often a web-based or mobile chat room—facilitates real-time conversation among students (and possibly the instructor) during a traditional lecture (Educause Learning Initiative, 2010), and "can entice students who rarely raise a hand to express themselves via a medium they find as natural as breathing" (Trip, 2011, para. 5). Previous research has shown backchannel use increases students’ feelings of connectedness in lecture and allows students to provide real-time feedback within a "micro discussion" context, similar to text messaging or micro-blogging (e.g., Twitter) (Aargard, Bowen, & Olesova, 2010). The backchannel affords audience engagement by reporting, enhancing, and commenting on course content via a public website and offers the presenter more direct and immediate feedback (Atkinson, 2009).
Audience response systems. Polling via an audience response system also has received much attention in the educational community. Put simply, "[audience] response systems are instructional technologies that allow instructors to rapidly collect and analyze student responses to questions posed during class" (Bruff, 2009, p. 1). ARS, also known as "clickers," have more recently come to include web-based polling using text messaging and mobile devices. Research about polling has shown that it yields more active student involvement, increased attention levels and interaction, and increased comfort in answering questions (Filer, 2010). Furthermore, while use of an ARS did not increase post-lecture quiz scores, students using an ARS reported significantly higher course satisfaction scores (Filer, 2010). ARS give instructors the ability to "assess students’ existing knowledge and competencies" as well as "provide immediate and appropriate feedback during instruction to help clarify student understanding of class content" (Kyei-Blankson, 2009, Conclusion, para. 4). Finally, an ARS can be used to acknowledge each individual’s response and draw attention to the diversity of responses, all while keeping responses anonymous and leaving instructors with quantitative information about student comprehension (Campt & Freeman, 2010).
Polling and the backchannel further provide both pedagogical and practical benefits in the classroom. For example, both technologies can be used to identify common errors or misconceptions and allow immediate instructor feedback by way of additional explanations or clarifying examples (Ambrose, Bridges, DiPietro, Lovett, & Norman, 2010). The anonymity provided by polling removes the "embarrassment factor" and gives students a safe space to answer questions honestly (Bruff, 2009), which promotes an engaging classroom environment (Ambrose et al., 2010). Moreover, the backchannel provides meaningful technology use by students during class. As Derek Bruff told The New York Times , "professors could reduce [tuning out, checking email, and online shopping] by giving students something class-related to do on their mobile devices" (Tripp, 2011, para. 10).
Together with the HyFlex model, the technologies mentioned above can be leveraged to attain the scalability needed for increased course enrollment, while also meeting the demand among current college students for autonomy and choice in the consumption of educational content (Dahlstrom, 2012). While this particular study was conducted in a general education Statistics course, the results and recommendations are geared for broader application in any academic discipline.
Method
Setting
A total of 161 undergraduate students enrolled in the pilot section of Statistics 145 (Autumn 2011) at a large, public Midwestern university participated in the study (98.8% of the course enrollment). Statistics 145 is an algebra-based introduction to data analysis, experimental design, sampling, and linear regression. The course satisfies the "data analysis" portion of the general education requirement for students in several programs across the university. For comparative purposes, the control group for this study consisted of the 168 undergraduate students enrolled in two additional sections of Statistics 145: one section with 114 students during the day, and one at night with 54 students (the pilot section was held during the day). At the time of this study, Statistics 145 was a 10-week course with three 48-minute lecture sections and two 48-minute recitation sections per week. As will be outlined, the pilot study focused almost completely on the lecture component of the course.
Following Beatty’s HyFlex model, the primary goal for this study was to provide students with attendance options. Specifically, the live attendance options made available to students in this HyFlex model were face-to-face (traditional) attendance and synchronous online attendance. The course was structured so students could choose how they wished to attend lecture on a daily basis; no requirements were imposed on how often students attended online or traditionally. For example, students could attend completely face-to-face, completely online, or some combination of the two. The only mandated face-to-face portions of the class were recitation periods and examinations.
Choice of technology. This study was initially motivated by a university-specific semester conversion, which caused a higher demand for large lecture spaces. Another primary goal was to serve more students with reduced access to large lecture spaces. The HyFlex model allowed the registrar to book a large space for two lecture days compared to the typical three. An implied third goal, consistent with the concept of HyFlex equivalency, was to standardize the students’ experience of the class as much as possible across the attendance methods (Beatty, 2010). Planning for and implementation of technology became a large component of the project, as a variety of tools were needed to enable students attending synchronously online to also fully participate in lecture meetings. To this end, a backchannel and an ARS were introduced; the specific technology chosen for this project was Poll Everywhere for both the ARS and backchannel. The authors based this choice on factors including technological compatibility, cost, and accessibility. The usage and implementation of Poll Everywhere will be described more fully; however, ongoing expansion of this research includes studying the viability of other technologies, such as Twitter and Top Hat Monocle.
Procedures. The non-pilot sections did not offer attendance choices, use of an ARS, or use of a backchannel, but aside from these differences the three sections of Statistics 145 referenced in this paper (one pilot section, two traditional/control sections) were highly similar. Three different lecturers taught the different sections, but the instructors used the same PowerPoint slides, provided students with the same note outlines, received common training from the course coordinator, and participated in weekly meetings to discuss pace and progress of the course. All sections were taught with the same course objectives, namely to introduce students to the process of doing statistics, collecting data, enabling the use of statistical tools for data presentation and analysis, and learning common statistical procedures and summaries. Homework assignments and exams were common to the three sections, and the learning activities administered in the small-group recitation time were identical. Some variability was introduced by way of the team of graduate teaching assistants who led the recitation sections.
Lecture attendance was not required in either the pilot or non-pilot sections, but recitation attendance was required in all cases and contributed to a participation component of students’ grades. An additional portion of each student’s participation grade came from a completion-based activity, and this activity was different for the pilot and control sections. As in previous terms, students in the control sections were required to complete and turn in a worksheet based on the day’s recitation activity, which was graded for completion. Students in the pilot section, on the other hand, were required to complete a "lecture review quiz" after each lecture and before the following recitation meeting. This was a three- to five-question quiz, administered on the course management website, and was designed to reiterate large-scale concepts from lecture. This assessment had two goals: to ensure that students experienced the lecture content before they attended recitation, and to track how students were attending lecture. Lecture review quizzes were graded for completion.
HyFlex Model Implementation
Figure 1 . Screenshot of Adobe Connect. This screen is what students see when attending synchronously; here, in solving a worked problem, the instructor�s writing appears in real time.
During each lecture, both Miller (course coordinator, lecturer and primary instructor) and graduate research associate Risser were present, with Miller teaching at the front of the classroom and Risser situated with a laptop among the students to assist with setup and backchannel management. The HyFlex model was introduced on the first day of lecture, during which attendance methods and technological implementations were outlined.
Figure 2 . Screenshot of the backchannel, using Poll Everywhere. Responses with "(Risser)" refer to commentary provided by the research associate and classroom assistant.
To make synchronous online attendance possible, both lecture slides and an audio feed were streamed live using Adobe Connect web conferencing software. Adobe Connect was available for the project thanks to a university license, and served the needs of the HyFlex model well. Lecture slides were projected in class via a tablet PC, and the instructor wore a lapel microphone to provide an audio feed. Students attending online watched a live view of the instructor’s computer screen (Figure 1), most often in full-screen presentation mode showing PowerPoint slides with real-time writing. As well as being streamed live, lectures were recorded and posted on the course management website to enable future viewing and reviewing.
A backchannel and an ARS were incorporated to allow interaction between the instructor and students attending class synchronously. Using Poll Everywhere, the ARS consisted of web-based polling, and student participation was made possible through use of text messaging or Internet-enabled mobile devices. The backchannel allowed students viewing online to submit a question at any time during the lecture. Submissions would then scroll down the public backchannel site much like a news or Twitter feed (Figure 2), visible to other students attending synchronously. The graduate research associate moderated the backchannel, either responding to questions directly (also through the backchannel) or repeating the question orally in class for the instructor to provide a more thorough response.
Figure 3 . Screenshot of three Poll Everywhere multiple-choice questions. The questions highlight how the poll allowed the instructor to address the majority of incorrect responses: in 3(a), the second choice is actually correct; in 3(b), the third choice is correct; in 3(c), the first choice is correct.
In general, the primary function of Poll Everywhere is to provide a forum for multiple choice, true/false, and short answer questions, which are administered and viewed online (see Figure 3). On average, between one and four questions were posed to students each lecture. The specific questions chosen were designed to mirror the pedagogical foundations referenced in the introduction; namely, to reveal common misconceptions held by students in the classroom and to subsequently allow students who answered incorrectly to see that they are not alone in their misunderstanding.
Incorporation of the backchannel and online polling also benefited students who chose to attend face-to-face. Both the backchannel and the online polls allowed contributions via any device (e.g., text message or web browser), and students were able to contribute to these forums regardless of attendance mode. Participation in both the poll questions and backchannel was neither encouraged nor discouraged. Student participation in the polls and backchannel was anonymous—a function of the constraints of Poll Everywhere, but a reality that the authors were content with.
Findings
Data collected from the pilot study included self-reported attendance information, grades, an end-of-term survey, and focus groups. Additionally, students enrolled in all sections of Statistics 145 for the Autumn 2011 term completed a pre-test and post-test to assess progress during the term.
Comparison to Non-Pilot Sections
The first research question involved an overall comparison of learning between students involved in the pilot section of Statistics 145 and students in one of the two other Statistics 145 sections offered during Autumn 2011. To address this question, student learning was measured by the START test, a standardized test which assesses students’ statistical reasoning. The 14-question START test was developed as a shorter alternative to the 40-item Comprehensive Assessment of Outcomes in a First Statistics Course (CAOS) test, developed by the ARTIST team at the University of Minnesota as part of a National Science Foundation grant (Delmas, Garfield, Ooms, & Chance, 2007). Students in both pilot and non-pilot sections of Statistics 145 were encouraged to take this test at the beginning and end of the course; the variable of interest was the difference between the post-class score and pre-class score. A total of 42.9% of the consenting students (69 of 161) in the pilot section took both pre- and post-tests, as did 67 of the students from the non-pilot sections; analysis for this first question involves these 136 students.
Define µ pilot to be the true difference between post-class score and pre-class score for students in the pilot section; define µ control to be the true difference between post-class score and pre-class score for students in the non-pilot sections. A Welch’s two-sample t -test for equality of means was not statistically significant, with t (132.8) = -1.0447, p = .2967, with a 95% confidence interval to be [-1.401, .431] ( see Table 1 ). With a p -value of .2967, we do not reject the null hypothesis of equality of means—the HyFlex model performs no differently from the traditional classroom model with respect to student learning. Note that because the confidence interval includes more negative values (and the estimate itself is negative), the scores in the control sections are estimated to be better than the scores from the pilot section, although this difference is small and not significant.
Comparison Within Pilot Section
First, a comparison was made between the two categories of "Live" attendance. For each grade category (homework average, midterm, final grade), define µ f2f to be the true average grade for students attending most often face-to-face and µ AC to be the true average grade for students attending most often synchronously via Adobe Connect. For each of these three grade categories, a two-sample test for equality of means was again not statistically significant, with p -values of .7312, .2724, and .3448, respectively. See Table 3 for the full results and confidence intervals . For each grade category, with large p -values we are not able to reject the null hypotheses of no difference in means and conclude that there is no significant difference in homework grades, midterm scores, and final course grades for the two secondary attendance groups of students. This conclusion is again expected based on the fact that each confidence interval includes zero; as before, note that each interval contains more positive values, which indicates that the face-to-face grades are higher, but not significantly so.
To determine whether attendance methods have an impact on student success in the course, we next turn to a within-pilot section comparison. The proxy for student success chosen for our analysis was course grades; specifically, the average from 12 homework assignments, the midterm exam score, and final course grade. Analysis considered each of these three categories separately. To formally address this question, students were placed in categories by primary and secondary attendance method (measured by attendance mode frequency, self-reported in daily lecture review quizzes; see Table 2 ). The data used to address this second research question come from the 161 students in the pilot section who allowed their grades to be used for analysis.
Because no difference could be detected between the two groups of students who most often attended lecture live, it is safe to consider these students a single group, which allows us to further make the same comparisons as above, now based on the two groups consisting of students in the primary categories of "Live" and "Recording." See Table 4 for complete results . Again, large p -values and the fact that confidence intervals contain zero allow us to conclude that there is no significant difference in homework grades, midterm scores, and final course grades for these two primary attendance groups.
Affective Responses from Students
Both the end-of-term survey and focus groups addressed students’ affective responses to the HyFlex classroom model and provided qualitative information on what they did and did not like about the new course structure and inclusion of more learning technology. The end-of-term survey was aligned with university course satisfaction surveys; specifically, students were asked 30 questions about the class, with responses solicited on a 4-point Likert scale (strongly agree, agree, disagree, strongly disagree). Most of the questions addressed the technology used in the course, including both the technologies used previously (e.g., online course management system and publisher-created software used for online homework) and the new "instructional technologies" (IT) implemented in the pilot (IT was defined to be any combination of tablet-based slides in PowerPoint, statistical applets, Poll Everywhere, and Adobe Connect).
Ease of use and perceived benefits. A total of 77 students in the pilot study (47.9%) responded to the end-of-term survey. In general, students found the IT easy to use and helpful to their learning. For example, 80% of students either "agreed" or "strongly agreed" that it was easy to access the online lectures through Adobe Connect (another 12% marked "not applicable"); approximately the same percentage found it easy to respond to a lecture poll. Another 95% "agreed" or "strongly agreed" that IT made the course materials more interesting and increased understanding of the course concepts, and more than 70% said technology increased their participation in class over what was expected coming into the course. A subset of the results is available in Tables 6 and 7 .
Desire for flexibility. The feedback from the focus groups contained much overlap with the comments shared in the end-of-term survey. The first major theme in the focus groups involved the attendance choices. Most students chose their daily attendance based on whatever worked with their schedule on that particular day. Weather conditions, other class responsibilities, extracurricular commitments, and the need for extra sleep motivated Adobe Connect attendance; a preference for consistency and face-to-face instruction motivated in-class attendance. Interestingly, different students favored different attendance methods. Some preferred face-to-face attendance to online synchronous attendance because of the increased interaction with the instructor, the ability to speak with her after class, and an overall more engaging experience. Others disagreed, saying online synchronous attendance was better than attending face-to-face, as they were able to stay more alert, focused, and comfortable outside of the classroom. One student commented, "I enjoyed being able to watch class on my laptop sitting in my bedroom with a cup of coffee, in my pajamas. I learn better when I am comfortable, especially with morning classes."
Figure 4 . Final question from the end-of-term survey; highlights a preference for lectures that involve technology.
Level of participation. Students overwhelmingly replied that instructional technology increased their participation in class. Having lecture recordings for review enabled one student to go back and re-listen to how the instructor explained a difficult topic—a remediation tool not previously available. ("I really loved how if I didn’t understand a lecture, I could just go back and re-watch the lecture since they are all online.")Some students found the course difficult but "manageable" thanks to the use of instructional technology; students found the in-class polls helpful, engaging, and a fun way to check comprehension without the pressure of getting an answer wrong. One student even commented that using Poll Everywhere gave the feeling of being in a small class, where problems are worked out on a chalkboard.
The last question on the end-of-term survey asked students to select how they would most likely attend lecture if given a choice (Figure 4). Only five percent of respondents indicated that they would prefer face-to-face lectures with no instructional technology, while an overwhelming majority (57%) would select face-to-face lectures with instructional technology. A slightly smaller proportion (38%) would prefer completely online lectures.
Discussion
Based on the research foundations in Section 1 and the results and analyses shared in Section 3, a strong argument has been made in favor of the HyFlex model, especially for large, on-campus courses. Given the current state of the post-secondary education system, the flavor of blended learning described in this paper has great potential in its application, as it reduces the pressures placed on brick and mortar universities by increased enrollment and rising popularity of online education. It is notable that after experiencing a 10-week HyFlex course, only five percent of respondents in the end-of-term survey said they would prefer a class that was fully face-to-face with limited instructional technology. The HyFlex model additionally satisfies the needs and desires of today’s college student by providing educational choices, opportunity for a student-driven curriculum, an engaging classroom atmosphere, and a concrete framework for incorporating the technology students are already using every day. For example, a 2010 Pew survey reported that 54% of all teens contact their friends daily using text messaging and 25% do the same with a social network site; only 33% of teenagers talk to their friends face-to-face on a daily basis (Lenhart, 2010). Furthermore, the HyFlex model accomplishes all of this while maintaining quality instruction and avoiding a decrease in course outcomes (operationalized by grades).
In the previous section, we discussed the ways in which this pilot study improved the quality of our course. The feedback received from students emphasized an increase in participation, increased interest in the course materials, and increased understanding of core concepts. The new components of this model were easy for students to use, from participating with in-class polls to accessing online homework assignments. Additionally, students were grateful for the availability of lecture recordings, which were posted shortly after class each day. Initially, the recordings were posted so that students who missed class could still experience the lecture content; however, we found that students were using the recordings as a remediation tool, going back through the videos to review difficult concepts. The impetus for increased participation in class came from a removal of the "embarrassment factor," made possible by anonymous responses to the web polls and backchannel. Students were more willing to take risks and commit to an answer even if they were not completely confident in their selection.
Technical Challenges and Self-Reported Attendance
However, our HyFlex curriculum model is by no means perfect, and several difficulties should be noted. The most common problem experienced involved technical difficulties which, while disruptive to the flow of lecture, were not overly detrimental. Additionally, as the term unfolded, it became obvious that collecting attendance data through self-reporting was less than ideal. On average, only 75-85% of the class completed lecture review quizzes for each session; however, each student filled out enough review quizzes over the course of the term to enable categorization by most common attendance method.
More seriously, anecdotal evidence indicated misreporting of attendance ("I know Dr. Miller said my responses won’t affect my grade, but I’m still not sure I feel comfortable honestly reporting how I attend class") in spite of repeated encouragement to honestly report attendance. Fortunately, although we cannot verify individual student responses, we are able to compare reported attendance numbers with actual data on synchronous Adobe Connect attendance and lecture recording views for nearly all class sessions ( see Table 5 ). The good news is that students were quite honest in reporting their synchronous Adobe Connect attendance: on average, true attendance actually exceeded reported attendance. On the other hand, reported views of the lecture recordings were much higher than the actual views. Thus, while the reported numbers of lecture recording views are suspect, the tallies for synchronous Adobe Connect attendance seem to be quite reliable. In any case, future research would do well to establish a more accurate and objective method for collecting attendance data.
Combined with subjective daily observation, the self-reported data seemed to indicate that the total number of students actually experiencing the lecture content was comparable to other terms. Nonetheless, an important question to address as this project moves forward is: what is the best way to motivate students to experience the lecture, either through face-to-face or online attendance? Measuring student engagement is something that might inform this question, but just how to do so was not addressed in this project.
Next Steps
Despite the success of this project, the issues encountered throughout the HyFlex pilot study leave many research questions for future investigation. For example, better ways to monitor attendance and develop the desired classroom community are needed. A more controlled research design might eliminate the option to view a recorded lecture, allowing all students to interact synchronously regardless of attendance method. Providing ways for students to feel socially connected with their classmates becomes particularly important within a HyFlex model where some students are not physically present in the classroom. Furthermore, the technologies used in this pilot study have been somewhat piecemeal, and we hope to find a more cohesive technology to both simplify setup and provide students with a more seamless experience. As technologies continue to advance, we look for ways to find the appropriate technologies for both our needs and the needs of our students.
Development of the Backchannel
Based on feedback from the end-of-term survey and personal pedagogical philosophy, another future goal of this project is continued development of the backchannel. The backchannel is not currently in its optimal form, which would be a real-time, in-class forum for students to ask questions, answer other students’ questions, and host relevant discussion on lecture content. Students are not currently required to bring a laptop to class, which would be necessary for the superlative backchannel. While Poll Everywhere and Adobe Connect will continue to be used in the short-term, additional exploration of available technologies is motivated by this problem, and finding a more all-encompassing technology (as mentioned above) would greatly improve the usefulness of the backchannel and contribute to feelings of classroom community.
Similarly, a better logistical arrangement for monitoring the backchannel needs to be developed. As described in Section 2, the current version of the HyFlex model requires two people to be present in the classroom: one to lecture, and one to monitor the technology. This is unrealistic; in future terms the instructor will be required to monitor the backchannel while lecturing. It remains to be seen if this additional responsibility will prove to be too great a distraction.
Continued research is required to continually improve this model. The results demonstrate some benefits, with little-to-no harm compared to the traditional lecture model. As HyFlex gains traction, it will be important to determine best practices to maintain quality while reducing physical and personnel resources associated with large courses. The authors invite others to join the conversation and to work to advance the HyFlex model.
References
Aargard, H., Bowen, K., & Olesova, L. (2010). Hotseat: Operating the backchannel in large lectures. Educause Quarterly , 33 (3). Retrieved from http://www.educause.edu/ero/article/hotseat-opening-backchannel-large-lectures
Abdous, M., & Yoshimure, M. (2010). Learner outcomes and satisfaction: A comparison of live
video-streamed instruction, satellite broadcast instruction, and face-to-face instruction.
Computers & Education
, 55(2), 733-741.
Ambrose, S. A., Bridges, M. W., DiPietro, M., Lovett, M.C., & Norman, M. K. (2010). How learning works: Seven research-based principles for smart teaching . San Francisco: Jossey-Bass.
Atkinson, C. (2009). The backchannel: How audiences are using Twitter and social media and changing presentations forever . Berkeley: New Riders Press.
Beatty, B. (2006). Designing the HyFlex world [PowerPoint slides] . Retrieved from http://itec.sfsu.edu/hyflex/beatty_hyflex_aect2006.ppt
Beatty, B. (2010). Hybrid courses with flexible participation – the HyFlex design . Draft v2.2. Retrieved from http://itec.sfsu.edu/hyflex/hyflex_course_design_theory_2.2.pdf
Bruff, D. (2009). Teaching with classroom response systems: Creating active learning environments . San Francisco: Jossey-Bass.
Buhagiar, T., & Potter, R. (2010). To stream or not to stream in a quantitative business course.
Journal of Instructional Pedagogies
, 3. Retrieved from
http://www.aabri.com /manuscripts/09417.pdf
Campt, D., & Freeman, M. (2009). Talk through the hand: Using audience response keypads to augment the facilitation of small group dialogue. International Journal of Public Participation , 3 (1), 80-107.
Dahlstrom, E. (2012). ECAR study of undergraduate students and information technology, 2012 (Research Report). Educause Center for Applied Research. Retrieved from http://www.educause.edu/library/resources/ecar-study-undergraduate-students-and-information-technology-2012
Delmas, R., Garfield, J., Ooms, A., & Chance, B. (2007). Assessing students’ conceptual understanding after a first course in statistics. Statistics Education Research Journal , 6 (2), 28-58.
Educause Learning Initiative (ELI) Publication. (2010). Seven things you should know about backchannel communication.
Educause Library.
Retrieved from
http://net.educause.edu/ir/library /pdf/ELI7057.pdf
Fernando, N. J. S., Cole, J. V., Tan, P. L., & Freitas, J. C. (2011, June).
Live Lecture Streaming
for Distributed Learning
. Paper presented at Scanning the horizons: Institutional research in a borderless world higher education institutional research network conference, Kingston University. Retrieved from
http://www.heir2011.org.uk/conference-papers/papers/Fernando%20HEIR2011.pdf
Filer, D. (2010). Everyone’s answering: using technology to increase classroom participation. Nursing Education Perspective , 31 (4), 247-50.
Knapp, L. G., Kelly-Reid, J. E., & Ginder, S. A. (2012). Enrollment in postsecondary institutions, fall 2010; Financial statistics, fiscal year 2010; and Graduation rates, selected cohorts, 2002-2007. U.S. Department of Education. Washington, DC: National Center for Education Statistics. Retrieved from http://nces.ed.gov/pubs2012/2012280.pdf
Kyei-Blankson, L. (2009). Enhancing student learning in a graduate research and statistics course with clickers. Educause Quarterly , 32 (4). Retrieved from http://www.educause.edu/ero/article/enhancing-student-learning-graduate-research-and-statistics-course-clickers
Lenhart, A. (2010). Text messaging becomes centerpiece communication. Pew Research Center Publications. Retrieved from http://pewresearch.org/pubs/1572/teens-cell-phones-text-messages
Sloan Consortium. (2012). Effective Practices Award Winners. http://sloanconsortium.org/ep_award_winners
Trip, G. (2011, May 12). Speaking up in class, silently, using social media. New York Times . Retrieved from http://www.nytimes.com/2011/05/13/education/13social.html?_r=2&pagewanted=all
Tables
Test for the equality of means between pilot and non-pilot sections |
||||
Classroom model |
n |
Mean difference in pre- and post-class START test scores |
95% CI for the difference in means |
p -value |
HyFlex (pilot) |
69 |
.754 |
[-1.401, .431] |
.2967 |
Traditional (control) |
67 |
1.239 |
Note: CI means confidence interval.
Testing the hypothesis of no difference in effect of classroom model (pilot versus traditional) on student learning. The large p -value indicates that classroom model had no significant effect on student learning.
Primary Attendance Category |
Secondary Attendance Category |
Category Description |
Total Students ( n ) |
Live |
Face to Face |
Students attending lecture in person. |
56 |
Adobe Connect |
Students attending lecture synchronously online. |
21 |
|
Recording |
Plans to Watch |
Students reporting their attendance before watching the recorded video. |
41 |
Watched Recording |
Students reporting their attendance after watching the recorded video. |
39 |
|
Did Not Attend |
Students neither attending lecture nor planning to watch the recorded video. |
4 |
Categorization of students in the pilot section, by the method students most often attended lecture.
Test for the equality of means between face-to-face and Adobe Connect attendees |
|||||
Grade Category |
Most frequent attendance method |
n |
Average grade |
95% CI for the difference in means (f2f – AC) |
p -value |
Homework (20 points total) |
Face to face |
56 |
17.21 |
[-1.418, 2.001] |
.7312 |
Adobe Connect |
21 |
16.919 |
|||
Midterm Exam (50 points total) |
Face to face |
56 |
42.728 |
[-1.476, 5.074] |
.2724 |
Adobe Connect |
21 |
40.929 |
|||
Final Course Grade |
Face to face |
56 |
81.854 |
[-4.223, 11.731] |
.3448 |
Adobe Connect |
21 |
78.1 |
Note : CI means confidence interval; f2f means Face to face; AC means Adobe Connect.
Within-section grade comparisons between secondary attendance groups of "Face to Face" and "Adobe Connect." Large p -values are desirable; this indicates that choice of attendance method does not negatively impact grades.
Test for the equality of means between Live and Recording attendees |
|||||
Grade Category |
Most frequent attendance method |
n |
Average grade |
95% CI for the difference in means
|
p -value |
Homework (20 points total) |
Live |
77 |
17.131 |
[-1.343, .635] |
.4806 |
Recording |
80 |
17.485 |
|||
Midterm Exam (50 points total) |
Live |
77 |
42.237 |
[-2.862, .823] |
.2761 |
Recording |
80 |
43.256 |
|||
Final Course Grade |
Live |
77 |
80.83 |
[-6.315, 2.167] |
.3354 |
Recording |
80 |
82.904 |
Note : CI means confidence interval.
Within-section grade comparisons, between primary attendance groups of "Live" and "Recording." Again, large p -values are desirable; this indicates that choice of attendance method does not negatively impact grades.
Class Date |
Actual recording views |
Reported recording views |
Difference (Reported minus actual) |
Actual Adobe Connect attendees |
Reported Adobe Connect attendees |
Difference (Reported minus actual) |
21-Sep |
36 |
N/A |
N/A |
N/A |
N/A |
N/A |
23-Sep |
49 |
N/A |
N/A |
28 |
N/A |
N/A |
26-Sep |
34 |
13 |
-21 |
36 |
25 |
-11 |
28-Sep |
42 |
35 |
-7 |
30 |
21 |
-9 |
30-Sep |
46 |
36 |
-10 |
26 |
21 |
-5 |
3-Oct |
41 |
34 |
-7 |
39 |
35 |
-4 |
5-Oct |
39 |
37 |
-2 |
32 |
26 |
-6 |
7-Oct |
50 |
57 |
7 |
43 |
24 |
-19 |
10-Oct |
45 |
35 |
-10 |
36 |
31 |
-5 |
12-Oct |
47 |
45 |
-2 |
30 |
24 |
-6 |
14-Oct |
48 |
66 |
18 |
27 |
25 |
-2 |
17-Oct |
39 |
27 |
-12 |
42 |
37 |
-5 |
21-Oct |
56 |
61 |
5 |
22 |
25 |
3 |
24-Oct |
54 |
63 |
9 |
22 |
23 |
1 |
26-Oct |
42 |
63 |
21 |
19 |
16 |
-3 |
28-Oct |
44 |
87 |
43 |
12 |
17 |
5 |
31-Oct |
50 |
70 |
20 |
18 |
17 |
-1 |
2-Nov |
47 |
51 |
4 |
28 |
20 |
-8 |
4-Nov |
54 |
69 |
15 |
16 |
17 |
1 |
7-Nov |
45 |
64 |
19 |
30 |
25 |
-5 |
9-Nov |
39 |
60 |
21 |
27 |
21 |
-6 |
14-Nov |
45 |
68 |
23 |
23 |
17 |
-6 |
16-Nov |
42 |
67 |
25 |
21 |
14 |
-7 |
18-Nov |
43 |
71 |
28 |
18 |
17 |
-1 |
21-Nov |
48 |
57 |
9 |
21 |
18 |
-3 |
23-Nov |
50 |
75 |
25 |
15 |
16 |
1 |
28-Nov |
36 |
52 |
16 |
20 |
16 |
-4 |
30-Nov |
37 |
61 |
24 |
22 |
14 |
-8 |
2-Dec |
35 |
76 |
41 |
16 |
15 |
-1 |
Actual (unique) recording views and synchronous Adobe Connect attendance compared with reported numbers. Note: Reported recording views change drastically in mid-October, which corresponds to the time of the midterm exam.
Question |
Strongly Agree or Agree |
Disagree or Strongly Disagree |
N/A |
It was easy for me to respond to a lecture poll using Poll Everywhere. |
83.1% |
11.7% |
5.2% |
It was easy for me to submit a question to the backchannel using Poll Everywhere. |
67.5% |
10.4% |
22.1% |
It was easy to access online lectures with Adobe Connect. |
81.8% |
6.5% |
11.7% |
It was easy to access lecture recordings in the university learning management system. |
84.4% |
3.9% |
11.7% |
End-of-term survey for Autumn 2011 pilot section; questions focusing on backchannel, polls, and live streaming. Response rate 47.2%.
Question |
Strongly Agree or Agree |
Strongly Disagree or Disagree |
Instructional technology (IT) helpedme prepare for recitations. |
88.3% |
11.7% |
IT helped make the materials and activities in the course more interesting. |
96.1% |
3.9% |
IT helped increase my interest in Statistics. |
71.4% |
28.6% |
IT helped me increase my understanding of the concepts in the course. |
94.7% |
5.3% |
IT was worth the time I spent using it. |
92.2% |
7.8% |
IT took away from my learning in the course. |
9.1% |
90.9% |
IT helped me increase my participation in class over what I would have expected coming into the course. |
71.4% |
28.6% |
IT in this course helped meet my preferred learning style. |
83.1% |
16.9% |
I would recommend this course to other students. |
97.4% |
2.6% |
End-of-term survey for Autumn 2011 pilot section; questions focusing on general use of instructional technology (IT), defined to be any combination of tablet-based slides in PowerPoint, statistical applets, Poll Everywhere, and Adobe Connect. Response rate 47.2%.