MERLOT Journal of Online Learning and Teaching

Vol. 3, No. 1, March 2007


 

Online Learning, Quality, and Illinois Community Colleges



Deborah L. G. Hutti
Associate Vice President for Educational Services
Lake Land College
Mattoon, Illinois  61938
Components
USA
dhutti@lakeland.cc.il.us

  

Abstract

 

In 2005, the Illinois Community Colleges Online (ILCCO) conducted a survey of faculty, staff, and students in order to identify the pressing issues surrounding quality, retention, and capacity building related to online learning.  Over a six month period, nearly one thousand individuals from seventeen Illinois community colleges provided data relevant to these three issues.  The following article focuses on the issue of quality and online learning, and the information obtained related to quality.

         

The data collection method included three different tools:  an electronic survey of faculty, staff, and students; a focus group that included faculty, staff, and students; and interviews with select faculty, staff, and students.  The results of the analysis of the data collected indicated that faculty, staff, and students agreed on the quality benchmarks that were most important and least important to online learning.

 

In addition, components for maintaining and improving the quality of online learning were identified.

 

Keywords: Online, Online Learning, Quality, Community College, Distance Learning

 

 

Online Learning in Illinois

Technology has impacted learning at an extraordinary pace, perhaps like no other innovation in previous years, decades, and centuries. According to Seagren and Watwood, (1996) “Distance education [specifically virtual classrooms] may have a greater impact on the nature of higher education than any innovation since the invention of the printing press” (p. 1). In 1998, data showed that there were approximately 1.7 million enrollments in all distance education courses in the United States with approximately 1.4 enrollments at the postsecondary level. Ten percent of all undergraduate students enrolled in at least one distance education course from 1999 to 2000 with two-thirds of all distance education enrollments via online learning (Sikora & Carroll, 2002). During the 2000-2001 academic year, there were approximately 3.1 million enrollments in distance at the postsecondary level (U.S. Department of Education, 2003). Amazingly, the concept of technology driven teaching and learning moved from unique and unusual to mainstream and ubiquitous within a very short period of time (Blumenstyk, 1998; Milliron, 2000; Milliron & Miles, 2000; Selingo, 1998).

According to the Illinois Board of Higher Education’s Fall Enrollment Survey, fall 2004 headcount enrollment was 801,548 which included all public and private, four year and two year institutions of higher learning within the state of Illinois (IBHE, 2005).  During that same time period, Illinois colleges and universities reported that of the 801,548 total enrollments, 80,165 were classified as online (IVC, 2005).  Illinois colleges and universities had offered 5,279 online course sections (IVC, 2005) which represented a 41% increase from the fall 2003 semester.  The tremendous growth of online enrollment within the state of Illinois during this time period can be considered substantial; however, the change from fall 2003 to fall 2004 represents only one year of a phenomenal pattern that has occurred over the past seven years.  Since 1999, fall online enrollments have increased from 5,887 in 1999 to over 80,165 in 2004 (IVC, 2005).  

The majority of the online enrollment during fall 2004 occurred within Illinois independent colleges and universities (34,075) followed by the Illinois public community colleges (33,405), and the Illinois public universities (12,685) (IVC, 2005).  Yet the vast majority of the online enrollment within Illinois independent colleges and universities were classified as out-of state with only 27% enrollment classified as in-state.  The converse was true for Illinois public universities and Illinois public community colleges with the majority of their enrollment (84% and 97% respectively) classified as in state. 

In January 2003, Washburn surveyed directors of online learning at 66 different Illinois institutions of higher education in order to assess their perceptions of online learning (IVC, 2003).   He found that most online courses used within these institutions (80%) were developed by individual faculty members and that 85% of the institutions planned on increasing their online courses and programs in the future.  There seems to be strong emphasis within Illinois higher education to continue to meet the growing demand for online learning by relying on faculty experts to create new courses and programs.

Online Learning and Illinois Community Colleges 

From recent enrollment data, it seems as if students at Illinois Community Colleges have embraced online learning.  Over a five year period, fall enrollment more than tripled, increasing from 9,403 in fall 2000 to 33,405 in fall 2004.   The fall 2004 online enrollment represented 9.2% of the total enrollment (363,204) at Illinois public community colleges during that same semester (IBHE, 2005).    

In order to support the rapid growth in online enrollment throughout the State of Illinois, a variety of support initiatives have emerged.  Several of these initiatives thrive due to support received at the state and federal level and they include the Illinois Community Colleges Online (ILCCO) [http://www.ilcco.net], the Illinois Online Conference (IOC) [http://www.ilonlineconf.org], the Illinois Virtual Campus (IVC) [http://www.ivc.illinois.edu], the Illinois Online Network (ION) [http://www.ion.illinois.edu], the Illinois Virtual High School (IVHS) [http://www.ivhs.org], the Illinois Century Network (ICN) [http://www.illinois.net], and the Illinois Digital Library (IDAL) [http://www.idal.illinois.edu/].  Each one of these initiatives provides a level of support for online learning at the community college level as well as support for online learning throughout all higher education in Illinois and beyond.   

Quality and Online Learning 

Quality, though one of the most nebulous and subjective educational concepts regardless of academic area, has been a key issue within higher education.  And as with any significant teaching and learning movement within higher education, as the enrollment has grown, the desire among higher education professionals to ensure quality within the online framework has grown as well.  In 2000, Phipps and Merisotis identified benchmarks for online learning success (Institute for Higher Education Policy [IHEP], 2000).  This IHEP publication, which provides a listing of 24 benchmarks considered to be “essential” to ensuring quality within online teaching and learning, is one of the most requested of all IHEP documents (http://www.ihep.org).  In fall of 2003, the 19 community colleges and senior institutions of higher learning throughout the state of Maryland received funding from the Fund for the Improvement of Post-Secondary Education to “create and implement a process to certify the quality of online courses and online components.”   Dubbed Quality Matters, (http://www.qualitymatters.org), this initiative is one of the most extensive, relevant, and current studies to have investigated quality as it relates to online learning (see http://www.qualitymatters.org/about_qm.htm). As part of the FIPSE grant supporting this initiative, the Quality Matters research team assembled a literature review tied to a Matrix of Review Standards that provides sound insight into the terms quality and online learning (see http://www.qualitymatters.org/resources.htm#literature).    

In 2002, the state of Connecticut concluded a study of student experiences in online learning throughout Connecticut institutions of higher education.  Upon completion of their study, they noted that faculty played an important role in the success of students in online environments, as well as flexibility, discipline, communication, and student support systems (Burton and Goldsmith, 2002). 

Maryland and Connecticut are certainly not the only states that have sought to ensure quality as it relates to online learning.  In December of 2005, Ohio produced its first annual report of the state of e-learning throughout its higher education system (Ohio Learning Network, 2005).  Ohio’s e-learning enrollment had grown to over 37,000 students in the fall of 2004 which represented an increase of approximately 10,000 students over the fall of 2003.  Upon completion of the data analysis, they concluded that further study in three areas (finance, new markets, and impact on learning) was essential in the future.  In addition, four Ohio institutions were selected for case-studies.  All four indicated that quality was a top concern.  They found no significant difference between face to face learning and online learning within the selected measured student outcomes that focused on quality (Ohio Learning Network, 2005).  Student respondents to the Ohio case study indicated that they assess the quality of an online course in the same way that they assess the quality of a face-to-face course.  They believe that the quality of the instructor is a key indicator of the quality of all courses, face to face as well as online (Ohio Learning Network, 2005). 

Online Learning, Quality, and http://www.ilcco.net) asked the Chief Academic Officer (CAO) at each of the 49 Illinois Community Colleges to identify the key research interests for their respective colleges in regards to online learning.  Over a six month period and after a great deal of discussion, the CAOs identified the following as the top three key interests:  a) Quality – in particular, the quality of online courses; b) Capacity – in particular, the capacity of Illinois community colleges to attract and enroll students in online courses, to increase the number and types of online offerings, and to recruit more faculty to design and teach online courses; and c) Retention – in particular, the retention of students in online courses.  Based on this feedback, ILCCO began the process of researching all three areas with a strong and determined focus on the issue of quality. 

The Purpose of the Study 

Online learning, whether a semester long credit class or a short non-credit professional development activity is a critical teaching and learning activity in the 21st century.  As the number of students involved in online learning increases, the expertise of those willing to lead such learning must increase as well.  In 1970, there were a nominal number of computers connected to the internet.  With little time elapsing between the Internet explosion and the present, the interim period available for intensive study has been quite brief.  The unparalleled growth in online learning as a whole has not allowed for adequate, thorough, and methodical research to be completed covering issues like quality, capacity, and retention.  As the popularity of online learning continues to increase, so, too, does the need for continued research concerning such issues, in particular quality.  However, since the birth of online learning within community colleges throughout the state of Illinois and because of the brief amount of time that online learning has been available, only nominal research regarding any aspect of quality as it relates to online learning had been collected or analyzed.   

The study began by focusing on the identification of benchmarks considered important by students, faculty, and staff when considering quality as it relates to online learning.  The study used (with permission) the 24 quality benchmarks identified by the Institute for Higher Education Policy as a springboard.   In addition, the investigation of quality was directed by the following questions: 

  1. What do students believe are important benchmarks of quality in online learning?
  2. What does faculty believe are important benchmarks of quality in online learning?
  3. What do staff members believe are important benchmarks of quality in online learning?

The Participants 

Of the 49 public community colleges in Illinois, 17 were selected for participation in this study.  Community colleges were selected based on their size, geographic location, enrollment, online learning enrollment, and willingness to participate.  Two distinct groups of community colleges were identified.  Group I consisted of 12 community colleges considered to be representative of all Illinois community colleges.  Group II consisted of nine community colleges, based on self-nomination, as having special programs for the development and offering of online learning.  There were four community colleges that were in both Groups I and II.   All participants in Group I completed an online survey.  All participants in Group II completed an online survey and were interviewed by telephone or participated in an on-campus focus group. 

All faculty members who were teaching online courses during the fall of 2003 and all staff members who were directly dedicated to supporting online courses at that time were asked to participate.  Finally, the student sample requested from each participating community college to complete an online survey was based on an institution’s total online enrollment in fall 2003.  Table 1 displays the percentage of students asked to participate in the study as determined by the institution’s total online enrollment in fall 2003: 

Table 1.  Guide Used to Determine an Institution’s Student Sample Size

 

 

Total Enrollment in Online Classes Fall 2003

Sample Size Percentage

 

 

500-1000

30%

1000-1500

20%

Over 1500

10%

 After incomplete and duplicate entries were eliminated, 1024 respondents provided valid responses to the survey distributed as follows:  a) 42 staff members, b) 177 faculty members, and c) 805 students. 

Respective to Group II, researchers contacted three faculty members and one staff member at each college location and conducted a 45 minute phone interview.  In addition, each college in Group II set up a focus group consisting of fall 2003 online students that met with researchers for approximately 90 minutes.  

The Instruments 

The Online Surveys 

As discussed earlier, the surveys developed for this study were based on the Institute for Higher Education Policy’s 24 Quality Benchmarks (IHEP, 2000) with a separate survey designed with appropriate wording for each group of participants (faculty, staff, and students).  All surveys, regardless of respective respondents, were housed and completed via the ILCCO website and all response data were maintained, collected, and managed by ILCCO’s webmaster.   All survey information was submitted anonymously in order to protect the confidentiality of the participants. To prevent duplicate submissions, each person asked to complete the survey received a clearance code that could only be used once to gain access to the survey instrument.  

The first section of each online survey asked respondents for demographic information which assisted researchers in developing an accurate picture of faculty, staff, and students involved in online learning.  The second section of the survey asked respondents questions regarding quality and online learning.  Using a four point Likert Scale (Strongly Agree = 4, Agree = 3, Disagree = 2, and Strongly Disagree = 1), respondents were asked to rank the importance of IHEP’s quality benchmarks.  In addition, using the same four point Likert Scale (Strongly Agree = 4, Agree = 3, Disagree = 2, and Strongly Disagree = 1), respondents were provided with an opportunity to indicate whether or not the quality benchmark was occurring at their respective institution.  The researchers selected one Illinois Community College who was not among the Group I or Group II participants to pilot test the online surveys.  Once comments were received, the researchers made appropriate changes accordingly. 

The Interviews 

In order to complete the faculty and staff interviews, the researchers developed a list of key interview questions. One set was developed for faculty and one set was developed for staff, with both sets being very similar allowing for differences when discussing areas relevant to job activities.   The key interview questions were used when speaking with interviewees via telephone.  The average length of time for the interviews was approximately 1 hour.  The list of nineteen questions included a few demographic questions in order to allow the researchers to describe the setting / framework of those interviewed.  

The Focus Groups  

The researchers developed a separate set of focus group questions that were used to guide the student focus groups.  The focus group questions were similar to the interview questions asked via telephone of the faculty and staff; however, the focus group questions were tailored to student activities.  The only demographic information collected was the number of students who attended the focus group session. 

All faculty, staff, and students who either participated in telephone interviews or participated in focus groups were also asked to complete the online survey.  In addition to the online surveys and to the interviews and focus groups, an environmental scan of online learning within three state systems (Colorado, New York, and Texas) was completed to help provide a frame of reference for data collected within Illinois. 

Procedures 

Online Survey Data Collection 

To begin the data collection process, the surveys were placed on an ILCCO website.  The CAOs were contacted and asked to direct faculty, staff, and students from their respective institution to the website.  The researchers sent instructions to the CAOs at the participating Illinois Community Colleges which provided details regarding the entire process and included a deadline for survey completion.   The ILCCO webmaster developed a data collection system that funneled response data received from the respondents through the website to a spreadsheet that could be manipulated during the data analysis process.  Completion of the online survey was on a volunteer basis, and all information received from the surveys was held confidential.   

Interview and Focus Group Data Collection 

To begin the interview and focus group process, the CAOs were contacted and asked to submit names and contact information for at least three faculty members teaching online courses during the spring of 2005 and at least one staff member who was providing direct support for online learning during the same semester to the interview team.  The CAOs were also asked to assemble a group of students who were enrolled in online courses during spring 2005 and who were willing to participate in a focus group session.  The researchers sent the list of interview and focus group questions to the CAOs for their review prior to the activities.   

Data Analysis 

Creswell (2003) described data analysis as an ongoing process with continual reflection, involving data that has been obtained by the researcher from the participants that had been tailored to the specific strategy used in the study.  Merriam (2001) defined it as “the process of making sense out of the data” (p. 178).  Best and Kahn (1998) discussed changes in data analysis respective to the advent of the use of the computer and acknowledged that the computer as “one of the most versatile and ingenious developments of the technological age” has significantly altered the way in which researchers analyze data.  It has simplified complicated research designs and the data analysis process (p. 428). 

The data analysis for the survey portion of this study included the reporting of the number of valid responses, or useable responses returned from the respondents, as well as the number of missing responses or the number of invalid responses.  The data analysis also included the frequency of each type of response, the percentage of the total responses, the percentage of the valid responses, and the cumulative percentages of the valid responses.  All relevant quantitative data collected were analyzed using means, standard deviations, medians, and modes.  All data were collected and placed in an Excel workbook, using separate spreadsheets to assist when necessary.   

Using the statistical calculation formula features within the Excel workbooks, all statistical calculations were calculated for all questions that had been assigned Likert scales.  The calculations were completed on an aggregate basis as well as an individual basis when necessary. 

Finally, the qualitative data provided by the respondents throughout the interviews and the focus groups were analyzed by transcribing and placing all information into searchable Excel tables and seeking repetition of words, phrases, themes, and concepts.  An attempt was made to identify comment characteristics that had been repeated not only by an individual respondent, but by multiple respondents in multiple interviews and focus groups.  The more words, phrases, themes, or concepts had been repeated, the greater importance they were presumed to play. 

Findings 

The Survey Respondents 

Students.   According to the demographic data collected via the surveys, 45.6% of the respondents were between 18 to 25 years of age.  That percentage mirrors the percentage of total students age 25 or younger (45.7%) enrolled during fiscal year 2004 within the Illinois Community College System (Illinois Community College Board, 2004).  The majority of the respondents were full-time students (57%) with a very even distribution of first year students (35%), second year students (30.7%) and students having completed at least four college semesters (33.4%).  Seventy-five percent of the respondents were employed.  Sixty percent of the employed respondents were employed full-time.  

In addition, 39.7% of the respondents were new to online learning, having enrolled in an online course for the first time.  Most respondents (53.6%) were enrolled in only one online course during spring 2005, with 27% enrolled in two online courses and 18.2% enrolled in 3-4 online courses.   An overwhelming 84.8% of respondents stated that they were using their home computers to complete their courses.  The remainder of the respondents identified using workplace computers (7.1%), school computers (5.1%), or computers located at other locations (3%).  The majority of the survey respondents stated that their technology expertise was at an advanced level (58%), with 31% stating that they were at an intermediate level and 10.2% at a novice level. 

Faculty.    According to the demographic data collected from the faculty respondents to the survey, overall they were a very experienced group.  Over 70 % of the faculty respondents had more than 10 years teaching experience (including K-12, community college, and university experience) with 15.4% having 6 to 10 years experience and 12.3% having 2-5 years experience.  Over 54% of the respondents had taught for more than 10 years in a community college setting, and 45.2% had taught more than 10 years at their current institution.  Not only did the respondents have strong experience levels, the majority (57%), albeit narrow, was full-time faculty.   Nearly 37% of the respondents taught between 12-24 credit hours face-to-face during the combined fall 2004 and spring 2005 academic semesters, with 24.8% having taught between 1-11 credit hours.  Furthermore, the respondents had a high level of experience teaching online with 57.6% having taught courses online between two to five years and 25% between six to ten years.  More than 60% had taught online for their current institution between two to five years, and 61% had taken an online course as a student.  The overwhelming majority of the faculty respondents rated their tech; The survey respondents categorized as online support staff provided similar demographic information as the faculty respondents.  Three fourths of the staff respondents had been in their current position for five years or less, with 17.5% in their position less than one year, 12.5% between 6-10 years and 12.5% more than ten years.  Nearly 58% had been employed at their current institution in some capacity for at least six years and 65% had spent at least six years employed in a community college setting.  Nearly 60% of the staff respondents indicated that no less than 40% of their workload was devoted to online learning.  Approximately one fourth of the staff respondents had online teaching experience and slightly more than 50% had been a student in an online course.  Three fourths of the staff respondents rated their technology expertise as advanced and one fourth rated it as intermediate.  

Participants Interviewed.  As all individuals who participated in the interview / focus groups were asked and expected to complete the online survey, the demographic information corresponded to that collected through the survey.  Some additional information was gathered that provides some additional details about the participants interviewed.  One faculty participant interviewed had taught over 100 online sections, and one was teaching online for the first time.  The faculty participants taught a wide variety of course including composition, history, humanities, mathematics, speech, psychology, science, word processing, automated office technology, management, accounting, criminal justice, nursing, and computer technology.   The staff supporting online learning did so in a variety of ways including instructional design, technical training, course management support, faculty website development, internal online committee participation, and institutional representation on statewide online learning initiatives.  Nearly 100% of the staff participants interviewed reported to the academic side of the college.   

The Quality Benchmarks 

Throughout the survey, all three groups of respondents were asked to rate the importance of each quality benchmark using a four point Likert Scale (Strongly Agree = 4, Agree = 3, Disagree = 2, and Strongly Disagree = 1).  Table 2 displays the benchmarks with the highest means based on all respondents (see Table 2).    Table 3 displays the benchmarks with the lowest means based on all respondents (see Table 3).   


Table 2. Quality
Benchmarks Rated Most Important Based on Highest Means – All Respondents

 Benchmark Mean

Technical assistance in course development is available to faculty

3.75

A college-wide system, such as Blackboard or WebCT, supports and facilitates
the online courses.

3.70

Faculty are encouraged to use technical assistance in course development

3.68

Faculty give constructive feedback on student assignments and to their questions

3.66

Faculty are assisted in the transition from classroom teaching to online instruction

3.65

The institution has a documented technology plan

3.63

Faculty give feedback to students in a timely manner

3.62

Before starting, students are advised about the course to determine if they have
access to the technology required by the course design

3.61

Before starting, student are aware of course objectives, concepts, ideas
and learning outcomes

3.60

The technology is reliable and failsafe.

3.60

 

Table 3. Quality Benchmarks Rated Least Important Based on Lowest Means – All Respondents

Benchmark

Mean

Instructor training and assistance, including peer mentoring, continues throughout the online course

3.47

The educational effectiveness of the online courses between institution is assessed through an institutional evaluation

3.47

Students are actively engaged in analysis, syntheses, and evaluation as part of their online course and program requirements

3.47

The educational effectiveness of the teaching/learning process is assessed through an institutional evaluation

3.46

The online courses address student learning styles

3.44

Students have access to a virtual library

3.39

Students are provided with online information and hands-on training on library resources

3.32

Data on enrollment, costs, and successful innovative uses of technology are used to evaluate program effectiveness

3.30

Faculty are encouraged and aided in sharing online courses between institutions

3.08

The data collected were also separated according to the distinct respondent groups of faculty, staff, and student.  Upon completing appropriate statistical analysis, none of the mean scores of the individual faculty, staff, and student groups were found to be different at a statistical significant level of .05.  In addition, the faculty and staff were grouped according to several other distinctions including: 1) faculty who had taken classes online and those who had not, 2) faculty who had taught fewer than six sections online and those who had taught six or more sections online, 3) those who identified their computer skills as novice, intermediate, and expert, and 4) staff who had taught online and those who had not.  Students were divided into various groups as well including: 1) those who were 25 years or younger and those who were 26 and older, 2) those who were employed full-time and those who were either employed part-time or not employed, 3) those who had taken two or fewer online courses, those who had taken three or four online courses, and those who had taken five or more, and 4) those who identified their computer skills as novice, intermediate, and expert.  For each of these groups, multivariate tests were completed in regards to the means of their responses and the quality benchmarks.  As a result of this analysis, no statistically significant differences were found between the means of any of the groups.  The lack of statistically significant differences reinforced the consistency among the groups as far as the ranking of the quality benchmarks. 

Following their rating of the importance of each quality benchmark, the respondents were asked to rate the extent to which the quality benchmarks were occurring at their respective institution. Again, the respondents used a four point Likert Scale (Strongly Agree = 4, Agree = 3, Disagree = 2, and Strongly Disagree = 1) to complete the survey.   Table 4 shows the means of the responses received regarding the extent to which they believed that the quality benchmark occurred at the respondents’ respective institutions.   

Again, upon completing appropriate statistical analysis, none of the mean scores of the individual faculty, staff, and student groups were found to be different at a statistical significant level of .05; however, it is notable that the faculty means were discernibly higher than those of students in regards to 1) student interaction with faculty and students, 2) the timeliness of faculty feedback, 3) the agreement on expectations for assignments, and 4) students’ access to technology assistance.  In addition, the faculty means were notably higher than those of the staff on:  1) the constructive quality of faculty feedback, 2) the timeliness of faculty feedback, 3) instruction in proper methods of research, 4) online information and training given on library resources, and 5) the assessment of teaching and learning. Finally, the means of the staff responses were higher than students on: 1) students’ access to virtual library, 2) students’ access to technical assistance, 3) timely responses from student services, and 4) the system of feedback about online courses. 

Table 4. Extent to which Quality Benchmark Occurs at Institution of Respondent based on Mean

 Benchmark

Mean

       Student

        Staff

      Faculty

Reliable and failsafe technology

3.33

3.30

3.15

College-wide computer system

3.66

3.70

3.68

Instructional materials reviewed

3.32

2.84

2.89

Courses reviewed periodically

3.35

2.82

2.90

Students engaged in analysis

3.26

3.11

3.14

Student interact with faculty and other students

3.36

3.43

3.28

Student interaction is facilitated through a variety of ways

3.32

3.39

3.48

Faculty give constructive feedback on student assignments

3.41

3.35

3.46

Faculty give feedback to students in a timely manner

3.34

3.30

3.45

Students instructed in proper methods of research

3.30

2.74

3.02

Before starting, students advised about the course to see if they are committed

3.40

3.18

3.01

Before starting, students advised about technology required by course

3.42

3.35

3.13

Before starting, students aware of course objectives, concepts, learning outcomes

3.41

3.35

3.31

Students have access to virtual library

3.12

3.25

3.19

Faculty and students agree on expectations for assignments completion and faculty responses

3.36

3.09

3.12

Online information about programs, admission, requirements, etc.

3.33

3.22

3.36

Students provided with information on library resources

2.99

2.75

2.87

Students have access to technical assistance

3.62

3.48

Student services answer students’ questions in a timely manner

3.20

3.36

3.26

Student services answer students’ questions accurately

3.23

3.28

3.26

Educational effectiveness of course is assessed through institutional evaluation

3.28

2.69

2.46

Teaching / learning is assessed

3.24

2.56

2.76

System for feedback about online courses

3.19

3.40

3.23

Online courses address learning styles

3.10

2.91

2.90

The Interviews 

At the conclusion of the interview and focus group activities, faculty, staff, and students consistently identified four components necessary for quality online learning programs including: 1) a strong administrative support team that collaborates with its faculty throughout the online learning process, 2) a system for continuous improvement of online teaching and learning, 3) reliable technology and reliable technical assistance, and 4) strong faculty, staff, and student online teaching and learning preparation / orientation programs. Throughout this study, the faculty, staff, and students were in nearly universal agreement that there was no significant difference between online teaching and learning.  In addition, the majority of the faculty interviewed stated that the outcomes between the face to face and online version of the same course were virtually identical.  Through the focus groups and interviews, the research team was able to extrapolate a list of components for maintaining and/or improving the quality of online courses (see Table 5). 

 

Table 5. List of components for maintaining / improving quality

             Components

1.

Use of a common course platform through a campus

2.

Availability of a well designed professional development program for faculty and staff preparing online teaching and learning activities

3.

Implementation of a readiness  / assessment program for students selecting online learning

4.

Use of online course assessment tools

5.

Adoption of clearly articulated standards and a list of components for all online courses

6.

Conveying a sense of personality by faculty and students throughout a course

7.

Use of an online course template – standardization

8.

Use of all available technology – low threshold and high threshold to enhance the learning environment

9.

Online mentoring for faculty and students

10.

Online tutoring

11.

Online readiness assessment for faculty and students

Conclusions 

Throughout the entire study, one conclusion that can be gleaned is that online learning is in a more mature phase.  The range of data collected during this study in particular throughout the interviews and focus group activities did not focus on the original forefront issues within online learning including those such as its academic credibility, the calculating of load relative to the online modality, the calculation of faculty compensation, and ownership issues.  Online learning has moved beyond the initial issues perhaps and hopefully because those issues have been resolved. 

Moreover, the results of this particular study demonstrate a remarkable cohesiveness among all respondents (faculty, staff, and students) in terms of their agreement on the quality benchmarks they found to be both most and least important.   Their unanimity provides credence to the idea that faculty, staff, and students all have common expectations in terms of quality and online learning even though their interaction, use, or activity within this modality is markedly different.  Because of this unanimity, it can also be concluded that there is a level of confidence concerning what constitutes or defines online quality among its daily users. 

Of the top ten quality benchmarks considered to be most important, six were directly related to technology, and of the nine quality benchmarks considered to be least important, only one was directly related to technology.   Similarly in terms of the extent to which the quality benchmarks were occurring at the respondent’s respective institution, four of the top seven strongest and two of the top seven weakest quality benchmarks occurring at institutions involved technology or technical assistance; however, the remaining three strongest and five weakest quality benchmarks did not.  Overall, the most important and those strongest occurring quality benchmarks at local institutions focused on technology.  The least important and those weakest occurring quality benchmarks did not.  In addition, six of the seven strongest occurring quality benchmarks at local institutions identified by the respondents appeared within the top ten benchmarks of greatest importance with the majority focusing on technology.  

Within the online learning modality, the technology used as the teaching and learning framework no doubt is paramount to its existence which may explain why benchmarks related to technology were considered most important and why they were occurring strongly at respondents’ respective institutions.  

Of greater interest is why quality benchmarks commonly associated with quality teaching and learning irrespective of learning modality were ranked of lesser importance and occurring to a lesser degree.  These findings seem to indicate that the technological segment of online learning overshadows the non-technological segment.  Issues such as a strong focus on learning outcomes, evaluation, assessment, educational effectiveness, and critical feedback should somehow be of greater value to teaching and learning than having technical assistance during course development.    If and when the teaching and learning move to the forefront and the technology becomes simply the offering mechanism, the measuring of the quality of online learning will be much easier to complete.        

References

Blumenstyk, G. (1998).  Leading community colleges go nationwide with new distance-learning network.  Chronicle of Higher Education, 44(44), A16-A17.

Burton, L. & Goldsmith, D. (2002).  Students’ experiences in online courses:  A study using asynchronous online focus groups. Connectifont-family:Arial">De A'Moreli, R. (1995). Exploring the Internet: Spectrum Universal.

Illinois Board of Higher Education. (2004). Data Book. Retrieved August 15, 2006, from the Illinois Board of Higher Education Website:  http://www.ibhe.state.il.us/Data%20Bank/DataBook/default.asp.

Illinois Community College Board. (2004).  Data and Characteristics.  Retrieved February 26, 2006, from the Illinois Community College Board Website: http://www.iccb.state.il.us/pdf/reports/databook2004.pdf

Illinois Virtual Campus. (2005).  Distance Education Enrollments at Illinois Colleges and Universities Summer/ 2005 Academic Year 2004-2005. Retrieved August 15, 2006, from the Illinois Virtual Campus Website:  http://www.ivc.illinois.edu/pubs/enrollment/Summer_05.html

Milliron, M.D. (2000).  The dotcommies are coming [Video Recording].  San Antonio: San Antonio College.

Milliron, M.D., & Miles, C. L. (2000).  Taking a big picture look @ technology, learning, and the community college.  Mission Viejo: League of Innovation in the Community College.

Moschovitis, C. J. P. (1999). History of the Internet : a chronology, 1843 to the present. Santa Barbara, Calif.: ABC-CLIO.

Ohio Learning Network.  (2005).  Bounded exuberance: e-Learning in Ohio.  Retrieved August 15, 2006 from, the Ohio Learning Network Website:  http://www.oln.org/about_oln/pdf/Bounded-web.pdf.

Phipps, R. & Merisotis, J.  (2000).  Quality on the line:  Benchmarks for success in internet-based distance education. Washington, D.C.:  The Institute for Higher Education Policy.

Seagren, A., & Watwood, B. (1996, February). The virtual classroom:  Great expectations.  Delivering graduate education by computer: A success story. Paper presented at the Annual International Conference of the National Community College Chair Academy.

Selingo, J. (1998).  Small, private colleges brace for competition from distance learning.  Chronicle of Higher Education, 44(34), A33-A35.

Sikora, A., & Carroll, C. D. (2002). A profile of participation in distance education: 1999-2000 (No. NCES 2003-154). Washington DC: Government Printing Office.

U.S. Department of Education. (2003). Distance education at degree-granting postsecondary education institutions 2000-2001 (No. NCES 2003-017). Washington, D.C.: National Center for Educational Statistics. 

 


 

Manuscript received 2 Nov 2006; revision received 5 Mar 2007.

 

This work is licensed under a

Creative Commons Attribution-NonCommercial-ShareAlike 2.5 License

 

 

 

 

 

 
Copyright © 2005 MERLOT. All Rights Reserved.
Portions Copyright by MERLOT Community Members. Used with Permission.
Questions? Email: jolteditor@merlot.org