Term paper quality of online vs. traditional students

Journal of Instructional Pedagogies

Term paper quality of online vs. traditional students

Stephen Hayward West Texas A&M University

Rex Pjesky West Texas A&M University ABSTRACT This paper uses a blind grading process to test if the performance of online students were different from traditional students using a term paper from an economics graduate course. Consistent with the literature, no significant difference was found between the scores of online students and those of traditional students. Also consistent with the literature, the only significant factors that influence student scores are variables that proxy for effort. Keywords: assessment, economics, graduate, online, traditional

Copyright statement: Authors retain the copyright to the manuscripts published in AABRI journals. Please see the AABRI Copyright Policy at .

Term paper quality, page 1

Journal of Instructional Pedagogies

INTRODUCTION

While the technology used and the format certainly has changed, distance learning has been a viable choice for students in the entire industrial era. The earliest form of what is now called distance learning were probably correspondence courses transmitted between teacher and pupil by the postal service. This slow means of communication between teacher and pupil has transformed into an environment where teacher and student can communicate costlessly with each other over virtually any distance instantly. While in the past, students doing distance learning would be isolated from each other, now they are not. Students in distance learning classes frequently communicate with one another as easily as they communicate with the professor (Gerlich, Mills and Sollosy, 2009).

This rapid change, which has mostly occurred within the last generation, has raised many new and old questions about how people learn and what the best relationship between professor and student is. To any engaged professor, and to any concerned student, an important question should be if the quality of the learning is equal across formats. It is possible that there is an ineffable quality to the teacher/student relationship and that technology can never replace this relationship. In contrast, maybe technology can make possible superior associations between teacher and student, and the academy is irresponsible if it does not embrace all the possibilities that technology presents (Bower, 2001).

To complicate the question, much of the demand for distance education does not stem from concerns of quality, but rather of convenience. Many students opt for online courses simply because their work or family obligations make taking classes in traditional formats difficult. Conversely, many students elect to pursue higher education in the traditional format because of the social experiences that being on campus are valuable to them. Regardless of the answers to these questions, it is important to make sure that the academy provides the best possible learning experience to students no matter what the mode of delivery is, online or traditional (Bower, 2001).

The purpose of this study is to estimate the difference in learning outcomes, as measured on the performance on a written paper, between students who enrolled in an online section of a class and students who enrolled in a traditional face-to-face section of the same class. The study's design assumes that variations in performance might come from variation in academic setting, in demographic characteristics, in academic ability, and in effort. Many, including Dutton, Dutton and Perry (2002) and Yukselturk and Bulut (2007) have taken this basic approach to assess the differences in online and in class modes of instruction.

The rest of the paper is organized as follows: First, there is an overview of the relevant literature to provide a context as to why this study is important and to provide a justification of the methods of this study. Then there is a discussion of some of the theoretical issues regarding the estimation of the data collected. Third, there is a description of the data and how it was collected. Fourth, an empirical model is developed to make inferences about the significance of the impact that mode of instruction and the other variables in the study have on the performance of the written paper. The final section will conclude the report and discuss implications.

LITERATURE REVIEW

There is a vast literature on distance learning. Much of this literature focuses on quality. The issue always has been that of quality. That is, those who advocate distance learning as a

Term paper quality, page 2

Journal of Instructional Pedagogies

viable alternative to the traditional classroom believe that the experience should be at least as academically rich as the experience that a student in a traditional classroom receives (Bower, 2001).

There are those that believe that the advantages of distance courses are so vast that the academy is doing a disservice to students in traditional course by not incorporating many of the tools and strategies used in the best practices of distance learning professors. Indeed, Turoff (1999) reports that if a professor takes an active role in facilitating the learning of the student (being the "guide on the side") and designs the course around a collaborative learning environment distance learning classes will have superior outcomes.

Yet most professors are trained in "hand to hand" teaching and indeed many of them were attracted to academics in the first place because they enjoyed direct interaction with students. The reality might be that the faculty member as the "sage on the stage." (Sanderson, Phua and Herda, 2000) The face-to-face instructor will use his/her intuition to gauge student reactions to create an effective classroom environment for learning. The technology of distance learning interrupts that feedback and perhaps denies a critical part of the learning process. Therefore, faculty might not just be luddites refusing to change for their own selfish reasons. Rather, they are simply refusing to follow the latest trend before they are presented with clearer evidence that the new technology helps students learn (Bower, 2001). Furthermore, it is not clear that, even though delivery options and technology are becoming more sophisticated and less costly, professors who deliver the courses or the information technology experts who deliver the machinery understand how people actually learn well enough to design a very effective learning environment that separates student from teacher (Greenagle, 2002).

Clearly, the question of quality in distance and online delivery of course material is an empirical one that can be tested. Terry (2007) used a large cross section of graduate students in business at a regional university to test for differences in learning (as measured by performance on class assignments) in campus classes, hybrid (those that use both online and in class instruction methods) classes, and online classes. Terry found that there was no significant difference in performance by students in campus and hybrid classes but students in online classes failed to perform as well on class assignments and did not meet course and program objectives as well as those students who had direct contact with professors.

DATA COLLECTION

The data for this report was collected from two sections of ECON 6306, a master's level macroeconomics course. One of the sections was purely online while the other was purely a "face to face" class. The classes were designed to minimize the difference between the sections.

To try to achieve equality of design between the online and traditional sections, several steps were taken. First, the classes had common objectives, textbooks and exams. The exams in the traditional class, like those in the online class, were "take home" tests consisting of essay and problem solving questions.

Both classes had weekly assignments that consisted of questions written by the professor that paced the students through the book. Students in the traditional section were held accountable for their performance on these questions by oral quizzing by the professor during class. Students in the online had the same questions, but they had to turn in written answers to the questions. Any relevant readings were distributed to both classes, and any substantial

Term paper quality, page 3

Journal of Instructional Pedagogies

discussions that occurred in the traditional class were also posted in message boards for the online class.

To maintain the integrity of the experiment, both online and traditional sections were given the same instructions, given the same background information on the topic. Students were to explore if the recent rounds of quantitative easing monetary policy by the world's central banks represented new, innovative policy actions by the central banks or were simply rehashes of old, traditional monetary policy. Economists were debating this topic in the press and in the blogosphere, so students in both sections were pointed to the same sources to use as references.

Finally, the papers were graded by the non-teaching coauthor of this study. The teaching coauthor numbered the papers so the grader would not know the name of the student or if the student was enrolled in the online section or the traditional section. This blind grading process was necessary since grading of paper is subjective and hence might be subjective to bias if the grader knew the name of the student or their online status.

Each paper on Quantitative Easing was given two grades, a traditional 100-point score and a separate critical writing score of ten points. The traditional score was broken down by content, summary, length and references with each category being awarded 40, 20, 20 and 20 points respectively. Since Quantitative Easing was a recent, hotly debated subject among economists as to its ability to work as a beneficial macroeconomic tool it allowed each student to take a position on whether it would work or not work thus giving them an opportunity to use critical writing. The ten points awarded for critical writing were divided into five points for stating a position clearly and five points if the position was defended.

METHODOGY

Davisson and Bonello (1976) proposed that student learning can be treated like a

production function in economics where production (i.e. student learning or performance) is a

function of class format or mode of instruction, the ability of the student, the effort of the

student, and the demographic characteristics of the student. That equation is:

Yi = F(Di, Ai, Ei, Vi)

(1)

Where

Y = performance on an assignment or learning outcome;

D = an indicator variable identifying different modes of instruction;

A = a student's inherent academic ability;

E = a student's effort;

V = a vector of demographic characteristics

i = an individual student in a cross section of students

Becker (1983) suggests a simple yet effective empirical model to estimate the relationship between student learning and vector of factors that might influence student learning. Equation (1) can be specified into an equation that can be estimated using data that is of particular interest to a specific research question. In the case of this report, we will estimate the model in equation (1) using the following equation:

Paperi = 0 + Onlinei1 + GMATi2 + GPAi3 + Overalli4 + Citizeni5 + MBAi6 +

Genderi7 + i

(2)

Term paper quality, page 4

Journal of Instructional Pedagogies

Where Paper = grade on the Quantitative Easing assignment; Online = 1 is student was in the online section, 0 otherwise; GPA = the student's cumulative graduate grade point average; GMAT = the student's score on the Graduate Management Aptitude Test; Overall = the student's final grade in the course exclusive of the paper grade; Citizen = 1 if the student is a US citizen, 0 otherwise; MBA = 1 if the student is an Master's in Business Administration student, 0 if the student is a Master's in Economics and Finance student; Gender = 1 is a student is female, 0 is a student is male; i = 44 students.

Equation (2) was estimated using a Tobit maximum likelihood technique. The Tobit estimator was selected instead of ordinary least squares since there was an upper bound on the dependent variable. The Tobit model presumes that if there is a discrete limit on the values of the right hand side variable, then some of the predicted values of the right hand side generated by the regression process may not exist. For example, it may be that one would expect students with a 3.5 GPA to score a 100 percent on an assessment test. This would imply that students with a higher GPA should score more than 100 percent of the assessment test. This, of course is not possible given that 100 percent correct is a practical upper limit on the assessment. Under these conditions, using OLS would not produce consistent estimates of the parameters. (Amemiya, 1985; Long, 1997) Five students in the sample of 44 received a grade of 100 on their paper, so a Tobit procedure is appropriate.

Table (1) presents the summary statistics. Fifty-two percent of the students were in the online class, 36% were US citizens, 64% were MBA students and 43 percent were female. The average score on the paper was 84%. The average grade of the student's tests and participation (That is, the average student grade exclusive of the paper) was 91%. The average GMAT was 484 and the average GPA was 3.64.

Discussion

The results of the estimation of Equation (2) are presented in Table (2). The most important result is the coefficient on the "ONLINE" variable. The coefficient is 5.82, which suggests that on average online students scored about 5.82% higher than traditional students do on the paper, holding the other factors constant. This coefficient however is not significant. The t-stat is 0.98, which suggests there is a high likelihood that this coefficient could happen by random chance in this sample.

Similarly, the coefficients on "GMAT," "CITIZEN," "MBA" and "GENDER" were insignificant. This is consistent with the literature that students' demographic characteristics generally do not affect student performance. The women in the sample scored 8.22 percentage points higher than the males in the sample, but with a t-stat of 1.87, the difference was not significant at the 5% level. There is little evidence in the literature that a student's innate ability as measured by achievement tests (the GMAT) have any predictive power on a student's performance on an individual class or assignment. Also, there is no evidence in the literature that a student's citizenship has predictive power on a student's performance on an individual class assignment. (Gerlich, et. al., 2009) A priori, the author's did not expect there to be any

Term paper quality, page 5

................
................

In order to avoid copyright disputes, this page is only a partial summary.

Google Online Preview   Download