• No results found

Examining demographics of MOOC learners : can embedded surveys improve or alter understandings?

N/A
N/A
Protected

Academic year: 2021

Share "Examining demographics of MOOC learners : can embedded surveys improve or alter understandings?"

Copied!
42
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

Graduate School of Child Development and Education

Examining demographics of MOOC learners:

Can embedded surveys improve or alter understandings?

Research Master Child Development and Education Research Master Thesis

Student: A.K.E. Van de Oudeweetering Supervisor: prof. dr. O. Agirdag

Reviewers: prof. dr. M.M.L.Volman & dr. A. Zand Scholten June 21, 2017

(2)

Abstract

Since Massive Open Online Courses (MOOCs) provide free, online education without enrollment restrictions, they have been expected to enhance educational opportunities among disadvantaged learners. In contrast, demographic data showed that disadvantaged individuals are underrepresented among MOOC learners and completers. However, as these findings are mainly based on email-based surveys with very low response rates, the threat of response error hampers explicit conclusions. This study examined whether including a survey embedded in the MOOC environment in addition to an email-based survey increases response rates, affects the representation of demographics, and influences the estimated associations between demographics and learning outcomes. An experiment was conducted in six MOOCs of the University of Amsterdam on platform Coursera. Learners (N = 3,834) in one cohort were randomly assigned to either receive a demographic survey by email or by the twofold approach. Results showed that the addition of the embedded survey caused response to increase from 6.9% to 61.5%. Regression models demonstrated that the odds of response were 23.97 times as high when learners received the additional survey. While the representation of demographics was not significantly affected by survey delivery mode, the effects of parental education and country of residence on learning outcomes was dependent on the delivery mode. The findings raise awareness on the importance of survey delivery and non-response and encourage further research on the representativeness of demographic surveys in MOOCs.

(3)

1. Introduction

The launch of the first, globally recognized Massive Open Online Course (MOOC) in 2008 announced new possibilities for the higher education sector. MOOCs are online educational programs that can integrate video-lectures, several types of assessments and discussion fora. MOOCs charge small or no tuition fees and allow people to enroll without any preconditions or restrictions. Due to these characteristics, MOOCs enable universities to disseminate extensive courses to very large numbers of students worldwide. To date, more than 550 institutions have facilitated MOOCs in various subject domains and languages (Shah, 2015). As an additional consequence, MOOCs provide opportunities for individual learners to access classes and receive certificates from any university around the world. Considering that MOOCs are less expensive, less selective and more time- and place flexible than traditional higher education, it has been expected that they could favor less privileged populations to engage in further learning (e.g., Kay, Reimann, Diebold, & Kummerfeld, 2013).

In contrast to the expectations, studies have shown that the majority of the MOOC population is male, well educated and from developed countries (e.g., Christensen et al., 2013). Moreover, there has been evidence to suggest that learners who are actually less educated or live in less developed countries are more likely to drop out of MOOCs early or to obtain lower grades (e.g., Dillahunt, Wang & Teasly, 2014; Greene, Oswald & Pomerantz, 2015; Kizilcec, Saltarelli, Reich, & Cohen, 2017). Notwithstanding, conclusions with regards to the enrolment and completion of underprivileged MOOC learners are mainly based on email-based surveys with low response rates (Van de Oudeweetering & Agirdag, 2017). For example, one of the most frequently cited studies to indicate the demographics of MOOC populations received response from 4.3 percent of the targeted population of learners (Christensen et al., 2013). The problem with such low response rates is that it decreases the likelihood of representative results (Fan & Yan, 2010). Especially since socially relevant

(4)

demographics such as income level, educational attainment and ethnic background could influence response to web-surveys (Couper, Kapteyn, Schonlau, & Winter, 2007), MOOC data that result from surveys with low response rates might misrepresent the actual population. In turn, the misrepresentation could distort the estimated effects of demographics on completion or other learning outcomes. Currently adopted strategies to enhance response and representativeness in MOOC surveys have not been ethically desirable, since these either obligated participation (Ho et al., 2015) or favored respondents over non-respondents in course-related activities (Gates, Wilkins, Conlon, Mossing, & Eftink, 2014). Therefore, it is not merely important to know to what extent the low response rates influence knowledge on social equality in participation and completion in MOOCs. More generally, it is essential to explore alternative strategies to increase the response rates and enhance representativeness in MOOCs surveys.

2. Background

2.1. Increasing response rates

In general, the rates of response are lower for surveys that are delivered online than for other survey deliveries (Fan & Yan, 2010). This is a pivotal concern for data validity, since the risk of biased results and the possible size of the bias increases with the proportion of non-response (e.g., Schouten, Cobben, & Bethlehem, 2009). Hence, increasing non-response is highly valuable for improving the quality and expected representativeness of web-survey data. Especially for MOOC surveys, as their response rates are generally very low, strategies to enhance response rates are useful for advancing the validity of the data.

There are two complementary theories that can help to understand how response rates can be raised. The first theoretical approach is the social-psychological theory, which assumes that survey response is an intuitive decision and is largely influenced by individual

(5)

dispositions (Porter, 2004). These dispositions, in turn, are shaped by societal developments, the research context or respondents’ characteristics and experiences (Groves, Cialdini, & Couper, 1992; Porter & Umbach, 2006). While this theoretical approach has been adopted to explain the relative underrepresentation of men and ethnic minorities in survey samples (Porter & Umbach, 2006), there is also empirical evidence on the impact of the research context as it is the only relevant aspect that can be manipulated. Following this line of research, it has been found that response can be increased when feelings of reciprocity, genuine authority and the relevance of survey response were emphasized in survey invitations (Porter, 2004; Fan & Yan, 2010).

The other theoretical strand in survey response, the social exchange theory, assumes that survey response is a rational decision that is based on cost-benefit considerations (Porter, 2004). Studies based on this theory mainly focused on providing benefits to respondents, especially financial in nature (e.g., Galesic & Bosnjak, 2009; Laguilles, William & Saunders, 2011). However, applying monetary rewards can be too costly in large samples and might even induce unethical stresses for those in underprivileged circumstances. Hence, minimizing the costs for respondents instead of raising the benefits might be a more appropriate approach in MOOC research. Considering that costs of survey response are mainly related to time and effort, reducing inconvenience in survey participation seems to be key for increasing response rates (Han, Albaum, Whiley, & Thirkell, 2009).

Jointly, these theories provide some suggestions how survey response in MOOCs can be stimulated. On the one hand, survey invitations should communicate some authority and a sense of reciprocity. As MOOC learners often have a high regard for MOOC instructors as research experts (Ross, Sinclair, Knox, & MacLeod, 2014), and some learners even experience a personal tie with their instructors (Ziegenfuss, 2016), response to MOOC surveys might increase when the survey is delivered by the course instructor rather than

(6)

through a platform-wide email. Furthermore, minimizing inconvenience in survey response is important. Email-based surveys might cause hindrance due the extended loading time of hyperlinks, the overburdening of email inboxes or the simple disappearance of emails in spam mailboxes (Fan & Yan, 2010; Reips, 2000; Sax, Gilmartin & Bryant, 2003). To avoid these setbacks, surveys among MOOC learners could be delivered in the online environment rather than through email. In this way, respondents do not have to open extra screens and the size and the content of the survey is directly visible to them.

2.2. Response errors

As enhancing response rates diminishes the likelihood of response errors, this means that demographic MOOC surveys with relatively low response might yield a different and perhaps a less accurate representation of the learning population than surveys with higher response rates. This issue of representativeness in MOOC surveys should not be confused with the issue of coverage errors in other web-surveys, which implies a biased sample because the survey only reaches those with Internet access and minimal digital literacies (Fan & Yan, 2010; Grandcolas, Rettie, & Marusenko, 2003). Since the MOOC population is bound to have Internet and some digital competences, this coverage error is not expected to apply in MOOC research. Instead, response error is a major impediment towards representative samples in MOOC surveys. This means that respondents share specific characteristics that are not generally shared among non-respondents, including demographics, which implies that the sample does not reflect the actual population (Grandcolas et al., 2003). For example, a study on another type of web-survey showed that even in a sample with Internet connections, selective response caused an overrepresentation of male, well-educated and young respondents from an ethnic majority (Bandilla, Couper, & Kaczmirek, 2014). There has been some empirical evidence to expect response errors in MOOC surveys as well. One study with

(7)

a purposive sample of regular learners, formal students and medical practitioners in a MOOC, found that survey response was highest among formal students (Annear et al., 2015). Likewise, a study in another MOOC indicated that response significantly differed between learners with particular language preferences in the course, including English, French, Bulgarian, Spanish, Greek and Slovenian (Colas, Sloep & Garreta-Domingo, 2016). As these studies suggest that socially relevant variables like age, gender, educational attainment and cultural background could affect response to web-surveys or MOOC surveys in specific, this suggests that there could be response errors in existing studies on MOOC learners’ demographics.

While the risk of biased results grows with the proportions of non-response, low response rates do not necessarily signify response errors (Schouten et al., 2009). Certain studies on sample representativeness in MOOCs found no or little differences between responders and non-responders in activity patterns and video lecture behaviors (Kizilcec, 2014; Shrader, Wu, Owens-Nicholson & Santa Ana, 2016). Although dissimilarities have been found in the results of a demographic MOOC survey with very low response rates (Christensen et al., 2013) and a survey with nearly complete response (Ho et al., 2015), these studies were not conducted in the same MOOCs or platform. Therefore, the differences cannot be attributed to response error. Statistical methods to estimate response errors or to adjust data to enhance representativeness neither suffice, since the accuracy of the resulting outcomes deteriorates with the magnitude of non-response and of the associated bias (Kizilcec, 2014; Schouten et al., 2009). For this reason, consistent research on strategies to increase response remains highly important for the validity of research results.

(8)

2.3. Estimated association between demographic background and completion

There is also a considerable body of research that predicted MOOC learning outcomes based on learners’ demographics. These studies generally suggested that well-educated learners from developed countries have an advantage in MOOC progress and completion (e.g., Dillahunt et al., 2014; Kizilcec et al., 2017). However, results about the effects of demographic background on MOOC learning outcomes are generally mixed and possibly confounded by the response errors in demographic surveys as well (Van de Oudeweetering & Agirdag, 2017). There are even grounded reasons to expect response errors in these findings. For example, several studies indicated that response to a survey was higher for those who complete a course than for those non-completing (e.g. Cisel, Bachelet & Bruillard, 2014; Rizzardini, Gütl, Chang, & Morales, 2014). In turn, response is lower for those who are less successful in assignments and quizzes (Gates et al., 2014). Response rates also appeared to be higher in undergraduate and graduate level courses in comparison to high school level courses (Kizilcec, Piech, & Schneider, 2013). As non-completers and less educated learners are less likely to be part of the monitored sample, the omission of their data might skew the estimated association between demographics and learning outcomes.

The accuracy of estimated effects on learning achievements can also be distorted due to the inclusion of larger numbers of MOOC enrollees that never access any assignments or exercises (Belanger & Thornton, 2013). That is to say, retaining these learners in the analyses might make it difficult to disentangle the effect of personal background on learning progress from the effect of differential intentions or motivations. In estimating the association between demographics and course completion, it therefore seems intelligible to focus on learners who have at least attempted assignments. In addition, there are differences in learning objectives and prior knowledge among active MOOC learners. Some learners start a course simply to learn about specific subtopics in the course or to try out a MOOC, which makes them less

(9)

inclined to finish or strive for excellent grades (Kizilcec et al., 2013). Moreover, it has been shown that cognitive abilities like prior subject knowledge, digital literacy and language proficiency are important to MOOC achievements (e.g., Banerjee & Duflo, 2014; Greene et al., 2015). It will therefore be relevant to consider whether and how learning objectives and cognitive skills might play a role in learners’ achievements and the estimated effects of their demographic background.

2.4. The present study

This study aims to examine whether delivering a survey in the MOOC environment in addition to the regular email-based survey will yield higher response rates and alternative outcomes than solely distributing email-based surveys. The assumption is that the additional survey will reinforce learners’ perception of the researchers’ authority and make survey response more convenient, specifically because it is presented in the MOOC environment. Therefore, this survey delivery may yield relatively more response and, by these means, reach another share of the targeted population. Knowledge on this issue can inform researchers in the field on superior strategies to deliver MOOC surveys. Furthermore, it could induce discussion among MOOC facilitators to improve the facilities for survey delivery and to check for representativeness. Finally, improved research tools could benefit knowledge on the participation and completion of disadvantaged versus advantaged learners and, consequently, on the status of social equality in MOOCs.

Based on these research purposes, the research questions in this study are:

1. To what extent does a demographic MOOC survey yield a higher response rate using a twofold survey delivery, which combines a survey embedded in the MOOC environment with an email-based survey, in comparison to using solely an email-based survey delivery?

(10)

2. To what extent does the twofold survey delivery affect the represented demographics in the survey sample, including age, gender, education, highest parental education, country of residence and ethnicity?

3. To what extent does the twofold delivery influence the estimated effect of education, parental education and country of residence on learning outcomes in a MOOC?

3a. To what extent does the twofold delivery influence the estimated effect of education, parental education and country of residence on learners’ completion of a MOOC?

3b. To what extent does the twofold delivery influence the estimated effect of education, parental education and country of residence on learners’ grade average in a MOOC?

Figure 1. Conceptual framework of the three research questions

3. Methods 3.1. Sample

The study focused on six MOOCs of the University of Amsterdam. In each MOOC, a sample was selected within a specific session, which is a cohort of four or eight weeks (see Table 1).

(11)

A share of the learners in these sessions was excluded from the sample, based on four criteria. First, as the study focused on learners who actually attempted to participate in the course, learners who did not watch the first video lecture were not included in the sample. Second, 55 learners who indicated to disagree with study participation through the informed consent were omitted from the sample. Although these disagreements could be counted as non-response, it was considered unethical to retain their data. Third, learners who did not receive the survey as planned in the experimental procedure were excluded. Finally, learners were only selected once. For responding learners, this meant that only the data for the course in which they responded to the survey earliest in time were included in the dataset. For non-responding learners who were enrolled in multiple courses, the data for the course where the experiment was implemented earliest in time were used. Hence, the final samples included unique learners who were correctly and randomly assigned to one of the experimental conditions, watched the first video lecture and who did not actively disagree with participation.

Table 1

Session dates and sample sizes for the selected MOOCs

Course Session Sample (N = 3,834)

Quantitative Methods Oct 24, 2016 – Dec 19, 2016 n = 665 Qualitative Research Methods Oct 24, 2016 – Dec 19, 2016 n = 327 Classical Sociological Theory Oct 31. 2016 – Dec 26, 2016 n = 377 Intr. To Communication Science Oct 31, 2016 – Nov 28, 2016 n = 499 Basic Statistics Dec 5, 2016 – Jan 30, 2017 n = 1795 Inferential Statistics Dec 5, 2016 – Jan 30, 2017 n = 171

3.2. Experiment

A randomized experiment was implemented in the six selected MOOCs. On Coursera, the platform that facilitated the dissemination of MOOCs for the University of Amsterdam, it was possible to implement an A/B test. This A/B test implied that the courses were separated in different branches, which are different versions of the course that are distributed simultaneously. Learners who were enrolled in one of the courses during the time of the

(12)

experiment were randomly assigned to one branch. Thus, they experienced one version of the course. For this study, one branch served as the control condition, while the other was the experimental condition. The only difference between the two branches was the survey delivery. Sample sizes for the control condition (n = 2,004) and experimental condition (n = 1,830) were different. Yet as learners were randomly assigned to the two conditions and were selected by the same criteria, these differences appeared due to chance and do not represent potential biases.

Control condition – email-based survey delivery In the control condition, learners received an email during the second week of the session. The email was signed by the course team and contained a short text to invite the learner to participate in the survey. Through a hyperlink in the email, learners could access the survey at an external survey website. The first page of the survey presented an information note on the goal of the study, the voluntary nature of participation, confidentiality of learners’ data and contact information of the supervising researcher. Learners could confirm their agreement or disagreement with study participation by choosing one of two response options (‘I have read and accept the terms presented above and I want to participate in the study’ or ‘I do not want to participate’). On the next page, the learners were presented with 17 questions about their demographic background, prior knowledge, their learning objective and reason for participating in the MOOC. In order to enhance participation, the learners received a reminder email 14 days later with a similar short text and the same link.

Experimental condition – twofold survey delivery In the experimental condition, the same two emails were sent out as in the control condition. In addition, the demographic survey was delivered as an ungraded quiz embedded in the course environment. This survey is referred to as the embedded survey. To maximize its visibility for the learners, this additional survey was presented prior to the first lecture. The embedded survey was

(13)

introduced with the same information note, consent question and included the same 17 questions on learners’ background as the email-based survey. This meant that the experimental condition enabled learners to respond to the same survey through two different channels. Therefore, the survey delivery in the experimental condition is labeled as the twofold survey delivery. Responses to the embedded as well as the email-based survey were recorded. In case a learner would use both survey deliveries, their response to the embedded survey would be used for the analyses.

3.3. Measures 3.3.1. Response rate

The response rate was estimated as the proportion of learners, included in the sample, who responded to the survey. These response rates were estimated per experimental condition. Learners in the sample who answered at least 50 percent of the questions in the survey were identified as responders, whereas learners in the sample who did not respond to the survey or answered less than 50 percent of the questions were considered to be non-responders. There are two reasons for this threshold. First, early attrition from the survey or meager response seems to reflect a lack of willingness to respond to the survey, which suggests that those with largely incomplete responses share characteristics with non-respondents. Hence, it would be most appropriate to consider them as non-respondents. Second, techniques to account for missing data or missing responses are less likely to yield valid results with increasing amounts of missing data (Graham, Olchowski & Gilreath, 2007). The 50 percent threshold is therefore intended to protect the relevance and accuracy of the results.

(14)

3.3.2. Represented demographics

A specific selection of demographic indicators was used to describe and estimate the represented demographics.

Gender was described by three categories, including male, female and other. A dummy variable with male as the reference category was composed. There were only few responders who indicated to not belong to the male or female category (n = 6). This raised concerns, since including such a small sample as a separate category could lead to a decrease in statistical power (VanVoorhis & Morgan, 2007). As it was considered unethical to randomly assign these six learners to either male or female categories, listwise deletion rather than multiple imputation of these six responses seemed the most appropriate option in the analyses regarding gender.

Age was based on respondents’ self-reported year of birth. To facilitate the analysis, age was estimated as the difference between year of birth and the year the analyses were conducted (2017).

Educational attainment was measured on a continuous scale with eleven response options. The variable was coded to represent values that ranged from 0 to 10 (0 = no schooling completed and 10 = doctorate degree). Although the use of this continuous scale might hinder a quick insight in the educational degrees of MOOC learners, it was deemed relevant to the main purpose of comparison in this study.

Highest parental education was measured using the same eleven response options as for educational attainment and recoded to the same continuous scale ranging from 0 to 10. However, the question was adjusted to target the educational attainment of the respondents’ mother as well as of the father. In addition, there was a response option for those who did not know the level of schooling of the parent. The highest parental education reflected responses for the parent with the highest level of education completed. In case the respondent did not

(15)

know the education level of one of their parents, the education level of the other parent was used for this variable. If the respondent did not know the education level of both parents, it was interpreted as a missing value. Through multiple imputation, plausible values for these missing responses were estimated.

Country of residence was estimated based on respondents’ self-reported country of residence. The responses were recoded to represent whether the learner lived in a developed country or developing country, based on the indicators of the UN Statistics Division (United Nations, 1999). A dummy variable with developed country as the reference category was composed.

Ethnic minority status was measured through subjective minority status. Respondents could indicate whether they considered themselves to be a member of an ethnic minority, or not, or whether they did not know. Responses for those who did not know the answer were recorded as missing values and later replaced by plausible values through multiple imputation. A dummy variable was composed, with non-minorities as the reference category.

3.3.3. Learning outcomes

Learning outcomes were assessed with two indicators. Data on both indicators were retrieved through data exports in Coursera. The first indicator was course completion, which reflected whether learners achieved a passing course grade or not (0 = no completion and 1 = completion). The other indicator, grade average, represented learners’ average grade over all assignments, quizzes and exams in the course, measured on a discrete continuous scale from 0 to 10.

(16)

3.3.5. Control variables

Several factors were anticipated to confound either all examined associations or only the estimated association between demographics and learning outcomes. To account for these factors, they were adopted as control variables.

Between-course differences Learners within the same course could share specific characteristics, since they were attracted to the same topic and interacted with the same course content. This means that learners within specific courses could share particular characteristics that are not shared among other learners. To control for these potential differences between courses, dummy variables for the six courses were included in the analyses with the course Quantitative Methods as the reference category. Estimating between-course differences as fixed effects was considered more appropriate than estimating them as random effects, as these differences were not the main interest of the research questions. Furthermore, the small number of courses and the large sample size within the courses made the estimation of random effects not particularly relevant (Snijders & Bosker, 2012).

Learning objective was measured through four response options, indicating whether the learner only wanted to watch lectures, wanted to complete some assignments, intended earn a certificate or did not decide on their learning objective yet. Since these response options do not reflect a clear ranked scale, and the objective to earn a certificate most accurately reflected the intention to complete the course, a dummy variable was composed. This dummy variable represented the objective to earn a certificate, while the other response options were integrated to represent the reference category.

Subject matter experience was measured through a self-report on a Likert-type four-point scale (0 = no experience and 3 = degree or job in the field). Hence, subject matter experience was adopted as a continuous variable that controlled for the positive influence of prior subject knowledge on MOOC learning outcomes.

(17)

ICT experience was measured on a Likert-type four-point scale as well (0 = no prior experience and 3 = advanced knowledge and degree or job in the field of ICT). By these means, ICT experience was used as a continuous variable to control for the potential positive impact of higher levels of digital literacies on learning outcomes.

English proficiency was an index of responses to three separate questions. Two questions targeted learners’ English proficiency in reading and writing, measured on a Likert-type five-point scale (0 = no proficiency and 4 = equivalent to native speaker). In addition, the frequency of speaking English was reported on a five-point scale (0 = never and 4 = every day). The average of the responses was estimated to compose the index for English proficiency. The Cronbach’s alpha = 0.78 of this index indicated that the internal consistency was satisfactory.

3.4. Data analysis

For the first research question, descriptive statistics on the response rates per survey delivery and per course were estimated to gain intuitive insight in the differences in response between the two survey deliveries. Furthermore, a logistic regression analysis was conducted to estimate the size and significance of the effect of the twofold survey delivery on likelihood of response. Survey response, as a dichotomous variable was regressed on a dummy variable for the twofold survey delivery and dummy variables that were used to control for between-course differences. Considering the potentially suppressing or mediating role of these control variables on the estimated effects, a model without the control variables was assessed and examined as well.

Since the second research question focused on the representation of demographics based on survey-generated responses, non-respondents were omitted from the analyses. First, descriptive statistics on the results of the demographic survey were estimated for the two

(18)

distinct survey deliveries to provide a general overview on the differences. Furthermore, regression analyses were conducted to single out the size and significance of the effect of the twofold survey delivery. Although the demographics were adopted as outcome variables, the results were not intended to reflect the effects on the demographics, as survey delivery is not expected to change learner characteristics. Instead, the outcomes reflect the effects on the representation of the demographics among the respondents in the sample. For demographics that were measured on a dichotomous scale, including gender, country of residence and ethnic minority status, logistic regressions were conducted. For the demographics that were measured on a continuous scale, including age, educational attainment and highest parental education, OLS regression analyses were conducted. Again, a dummy variable representing the twofold survey delivery was included as the predictor and the models were assessed with and without dummy variables for courses to account for the between-course differences.

The third research question consisted of two sub-questions. Since they both focused on represented demographics, non-respondents were omitted from the analyses. For the first sub-question, focused on course completion, learners who had not attempted any graded activity (e.g., quiz, peer-graded assessment) were excluded from the analyses as well. It was considered that the inclusion of the large amount of learners who have no intention to earn course credits could cause a misrepresentation of the effect of demographics. For the second sub-question, which focused on grade average, learners who had not completed the course were excluded from the analyses. This was done to avoid an overlap with the outcomes for course completion and to avoid a positively skewed distribution for course grade, which could impede the analyses. For both sub-questions separately, seven models were assessed. First, the models with mean-centered values for educational attainment, parental education and the dummies country of residence and survey delivery as predictors were investigated. Following, three interaction terms were included one-by-one and with and without control variables that

(19)

represented learning objective, prior knowledge and between-course differences. In this way, the distinct interaction effects could be assessed without controlling for other interaction effects. Moreover, the potentially confounding effects as well as the suppressing or mediating effects of the control variables could be examined.

While only respondents were included in the analyses for the second and third research question, there were some omitted responses or blank responses that represented the response option ‘I don’t know’. Multiple imputation was used to account for these missing values. With multiple imputation, each missing value is replaced by a value that appears plausible based on all non-missing data in the sample. This procedure is followed m times, so that m imputed datasets can then be pooled to provide a tenable prediction of the actual values of the missing data (Graham et al., 2007). For this study, five (m = 5) imputations were conducted. Research has shown that this is an appropriate amount of imputations to secure statistical power and unbiased results, given < 8% of missing data for all variables (Graham et al., 2007). The MICE package in R, which is built under R version 3.3.2., was adopted as it can flexibly impute continuous, binary or categorical data (Van Buuren & Groothuis-Oudshoorn, 2011). As multiple imputation is based on all non-missing information in the dataset, and the likelihood of accurate results is larger with more informative data, responses to all items in the survey were included in the imputation. Imputation was not necessary for the analyses for the first research question, since they did not include missing data. For the other (sub)questions, separate rounds of m = 5 imputations were executed for the three distinct samples. Descriptive statistics for the different samples are presented in Table 2.

(20)

Table 2

Relevant descriptive statistics for research question 2, 3a and 3b

Range M (SD) Question 2 (N = 1,263) M (SD) Question 3a (N = 801) M (SD) Question 3b (N = 163)

Twofold survey delivery 1/0 89.1% 86.4% 82.8%

Female 1/0 48.6% Ethnic minority 1/0 17.5% Developing country 1/0 41.7% 40.0% 38.7% Age 16-82 34.45 (11.44) Educational attainment 0-10 7.31 (1.83) 7.21 (1.92) 7.12 (1.84) Highest parental education 0-10 5.95 (2.67) 6.04 (2.66) 5.98 (2.68) Learning objective 1/0 23.0% 54.6% Subject experience 0-3 .82 (.75) .88 (.75) ICT experience 0-3 1.64 (.85) 1.73 (.82) English proficiency 0-4 3.40 (.67) 3.42 (.62) Completion 1/0 20.3% Grade average 0-10 9.01(.56)

Note. Descriptive statistics based on imputed data sets. For each variable, < 8% is imputed.

4. Results

4.1. The influence of survey delivery on response

Following the first research question, it was examined whether the twofold survey delivery generated a higher response rate than the email-based survey delivery. Descriptive statistics showed that the email-based survey attained a response rate of 6.9 percent (n = 138). The twofold survey delivery, which combined the email-based survey with a survey in the MOOC environment, reached a response rate of 61.5 percent (n = 1,125). This means that adding the embedded survey to the email-based survey substantially increased the response rate by approximately a factor of ten. Particularly, 5.9 percent of the learners assigned to experimental condition responded to the email survey, which is a similar percentage as in the control condition. In contrast, 59.2 percent of the learners in the twofold survey delivery condition responded to the embedded survey and 3.7 percent filled out the survey in both deliveries. This means that the largest proportion of response to the twofold survey is achieved through the embedded survey.

(21)

Furthermore, two logistic regression models that predicted response by survey delivery were assessed: one that controlled for course differences and one that did not (see Appendix A). This model with control variables explained significantly more variance in the outcome than a null model, χ2(6) = 1507.3, p <.001, Nagelkerke R2 = 0.45, and yielded accurate predictions for 71.8 % of the sample. Hence, the model was considered appropriate for explaining survey response. The effect of the twofold survey delivery, B = 3.18, p < .001, deviated positively and significantly from zero, controlling for course-level differences. More specifically, the odds of survey response versus non- response were 23.95 times as large for those who received both the twofold survey in comparison to those who only received the email-based survey. This substantiated that the additional embedded survey increased the survey response considerably.

4.2. The influence of survey delivery on represented demographics

To gain a global understanding of the potential impact of the survey delivery mode on the representation of demographics, differences between the two survey deliveries in the descriptive results of the demographic survey were examined (see Table 3).

Table 3

Descriptive statistics on demographics per survey delivery Range M (SD) Email-based survey (N = 138) Twofold survey (N = 1,125) Female 1/0 41.3% 49.5% Ethnic minority 1/0 13.0% 18.0% Developing country 1/0 39.1% 42.0% Age 16-75 38.01 (12.80) 34.01 (11.19) Educational attainment 0-10 7.43 (1.60) 7.30 (1.86) Parental education 0-10 6.27 (2.76) 5.91 (2.65) Note. Statistics are based on imputed data sets. For each variable, < 8% is imputed

(22)

The sample based on the twofold survey delivery represented a larger proportion of women, larger proportion of ethnic minorities and slightly larger proportion of learners from developing countries. Furthermore, the descriptive results of the twofold survey delivery reflected a lower average age, a slightly lower average educational attainment and lower parental education than the results of on the email-based survey.

The effects of survey delivery were analyzed through separate regression models for each outcome, with and without dummy predictors to control for between-course differences (see Appendix B and C). Evaluation statistics for the logistic regression models and assumptions for the linear regressions were checked before interpreting the results. For the logistic regression models, χ2 tests indicated that the models with control variables for between-course differences explained significantly more variance in the outcome than a null model. However, Nagelkerke R2 indicated that the percentages of explained variance

remained fairly low. For the linear regressions, the assumption of linearity could not be rejected in each of the models. However, the residuals for each model were not normally distributed, with high levels of skewness and high levels of leptokurtosis. This suggests that the values for the dependent variables were considerably invariant, which could cause an attenuation of the power to detect effects (Stevens, 2009). Furthermore, for age and educational attainment, the variances were not equal across groups. Since the larger variances were associated with the smaller group, this could cause an inflation of significance levels of the effects (Stevens, 2009). This means that the coefficients should be interpreted with caution. Table 4 presents the results of the models, each estimating the effects of the twofold survey delivery on the representation of a specific demographic, controlling for the differences between courses. For the logistic regression models, the intercept reflects the likelihood that learners in Quantitative methods identify with this demographic, while for the

(23)

linear regression models, these intercepts reflect the average value on the specified demographic in this course.

As reflected in the models, the effect of survey delivery on the represented average age, B = −3.59, p < .001, deviated significantly from zero. This indicates that learners who responded to the twofold survey delivery were, on average, 3.59 years younger than learners who responded to the email-based survey, taking into account between course differences. For the representation of other demographics, however, the effect of survey delivery did not appear to be significant. As the effects were neither significant for the models that did not control for between-course differences (see Appendix B and C), it was recognized that the control variables did not suppress the effects of survey delivery. Withal, this means that the differences in represented demographics between the two survey deliveries that were found in the descriptive statistics cannot be generalized to a larger MOOC population.

Although it was not the focus of the research questions, it was remarkable to find significant between-course differences in the represented learner demographics across courses. For example, the odds of response from a female learner or a learner from a developing country differed significantly between some courses (see Appendix B). As the effect of survey delivery is taken into account in these models, this seems to suggest that the representation of demographic characteristics could be mainly attributed to the courses. Logistic and OLS regression models for represented demographics (N = 1,263)

Female Ethnic minority

Developing

country Age Education

Parental education B SE B SE B SE B SE B SE B SE Intercept −.71** .21 −2.07** .32 −.63** .20 38.30** 1.15 7.72** .18 6.33** .26 Twofold survey delivery .29 .19 .40 .28 .12 .19 −3.59** 1.05 −.11 .16 −.36 .24 Nagelkerke R2 .06 .01 .01 R2 .02 .06 <.01

Notes. Fixed effects for between course-differences are included in the models, but not presented

(24)

4.3. The influence of survey delivery on the estimated association between demographic background and learning outcomes

In accordance with the third research question, it was investigated whether the survey delivery mode influenced the estimated effect of educational attainment, parental education and country of residence on two learning outcomes. For course completion and grade average respectively, logistic and OLS regression models with and without control variables were assessed (see Appendix D and E). The intercepts reflect either the likelihood of completion or mean values for the grade average of learners in the course Quantitative Methods.

4.3.1. Course completion

For each of the three models explaining course completion, assumptions were investigated. Indications of multicollinearity (VIF > 2) appeared for the interaction effects and the independent effects of each demographic variable when the relevant interaction term was entered in the model. This indicated that there was a considerable overlap between the effects of the demographics in the results of both survey deliveries. With regards to model fit, χ2 tests indicated that the models with control variables explained significantly more variance in the outcome than a null model (see Table 5). Still, Nagelkerke R2 reflected that the

percentage of explained variance remained quite small. Without control variables, the models did not explain a significant share of variance in the outcome. This seems to suggest that the control variables were more relevant for predicting course completion than the demographic indicators. Hence, these findings contravened the expectation that the selected demographic background characteristics would impact learners’ completion.

(25)

Table 5

Logistic and OLS regression models explaining course completion and grade average

Course completion (N = 801) Grade average (N = 163) Interaction Education Interaction Parent Education Interaction Country Interaction Education Interaction Parent Education Interaction Country B SE B SE B SE B SE B SE B SE Intercept −2.25** .67 −2.24** .68 −2.55** .71 8.83** .28 8.85** .28 9.07** .30 Education −.03 .14 <.01 .05 <.01 .05 .06 .07 −.02 .02 −.03 .02 Parent education −.01 .03 −.16* .08 −.01 .03 <.01 .01 .02 .03 <.01 .01 Country of residence −.20 .20 −.21 .21 .50 .48 −.07 .08 −.07 .08 −.44* .19 Twofold Survey delivery −.19 .25 −.18 .26 .15 .34 −.21 .10 −.25** .10 −.45** .14

Interaction terms Education x Twofold survey .03 .15 −.09 .07 Parent education x Twofold survey .18* .09 −.03 .04 Country residence x Twofold survey −.83 .52 .45* .21 Model evaluation -2 log likelihood 753.1 749.2 750.5 χ2diff 56.2** 60.1** 58.8** Nagelkerke R2 .10 .11 .10 R2 .37 .37 .39 Adjusted R2 .32 .31 .33

Notes. Fixed effects for control variables learning objective, subject experience, ICT experience, English proficiency and course differences are included

in the models, but are not presented. * p < .05, **p < .001

(26)

The coefficients for the demographic variables and the interaction terms in the models with control variables were specifically investigated to answer the research question, since these models showed to provide a more accurate prediction of course completion. For education, the independent effect and the interaction effect with the twofold survey delivery were both insignificant and practically negligible. This suggested that education did not particularly affect course completion, and that the alternative survey delivery would not provide a different conclusion. For parental education, the interaction term with the twofold survey delivery, B = .18, p < .05, was significantly larger than zero. In turn, the independent effect of parental education was significant in a negative direction, B = −.16, p < .05. This meant that the email-based survey would estimate a negative association between parental education and the likelihood of completion, controlling for other variables in the model. However, this association would be practically negligible if the results were based on the twofold survey delivery. Hence, survey delivery appeared to play a considerable role in the estimation of this association. For the final predictor country, the independent effect and its interaction effect with the twofold survey delivery were practically large, yet insignificant. The main explanation for the insignificance is the relatively large variance for both coefficients. Although the effects can therefore not be generalized, the size and negative direction of the interaction effect suggests that the twofold survey delivery might reflect a smaller likelihood of completion for learners from developing countries than the email-based survey.

4.3.2. Grade average

For the models predicting grade average, assumptions for linear regression models were checked. For the interaction terms and the associated demographic variables, there were concerns for multicollinearity (VIF > 2) as well. Only for the models with the control

(27)

variables included, assumptions of linearity and homogeneity of variance were satisfied. In addition, the models with the control variables explained more variance in the outcome than the models without the control variables. Again, this suggests that the selected control variables may be more relevant for explaining learning outcomes than the demographic indicators. However, the assumption of normality could not be satisfied in the models with control variables, due to high levels of kurtosis. This was an indication that the power to detect effects could be attenuated (Stevens, 2009).

Outcomes of the models demonstrated that the independent effects of educational attainment and parental education on grade average, as well as their interactions with survey delivery, were insignificant and negligible (see Table 5). This meant that neither education nor parental education seemed a relevant predictor for average grades, and that survey delivery did not alter the representation of this effect. However, the interaction term between country of residence and survey delivery, B = .48, p < .05, was significant and substantial. It demonstrated that, controlling for other variables in the model, the estimated average grade for learners from developing countries would be .48 points higher if only the twofold survey was used. Another important finding was that the independent effect of survey delivery was significantly negative. This indicated that the twofold survey delivery, in comparison to the email-based survey, reached learners who received on average lower grades.

5. Discussion 5.1. Key findings

Current knowledge on the demographic background of MOOC learners has mainly been acquired through email-based surveys that received very low response rates. Given that high rates of non-response indicate a high risk of unrepresentative results, present understandings about MOOC learners’ demographics and how these are associated with MOOC learning

(28)

outcomes may be inaccurate. Given the general inefficiency of email-based surveys to ensure response, the first purpose of this study was to examine to what extent an alternative delivery for demographic surveys in MOOCs could yield higher response rates. Furthermore, the aim was to investigate to what extent the survey delivery had an impact on the representation of MOOC learners’ demographics and on the estimated association between demographics and learning outcomes. Results of a twofold survey delivery, which combined the email-based survey with a survey embedded in the MOOC environment, were compared against the results of an email-based survey in a randomized experiment.

The results indicate that the twofold survey delivery yields a substantially and significantly higher response rate than the email-based survey. Whereas the email-based survey generated a 6.9 percent response rate in the selected sample, the twofold survey delivery respectively received 61.5 percent response. The higher response rate was attributable to the relatively large proportion of response to the survey that was embedded in the MOOC environment. With regards to theoretical assumptions, this could indicate that the survey in the learning environment might enhance feelings of reciprocity and authority. This would mean that the social psychological theories towards survey response are substantiated (Porter & Umbach, 2006). The results could also reflect the relevance of the social exchange theory to survey response. That is, a survey in the learning environment might minimize inconvenience associated with survey participation, which could make the relative costs of response less weighty in the learners’ cost-benefit evaluation concerning survey participation (Porter, 2004). In practical terms, the substantial increase in survey response supports the suggestion that the use of the twofold survey delivery, specifically the embedded survey, is a helpful strategy to raise response rates. Consequently, this means that this survey delivery could increase the likelihood of valid results in MOOC surveys without violating the voluntary nature of survey participation.

(29)

Regression analyses showed that the twofold survey delivery only yielded significantly different results with regards to the age of MOOC learners, with the average age being three years lower than in the email-based survey. Although the twofold survey represented somewhat larger proportions of ethnic minorities and women and depicted a slightly lower average educational attainment and parental education, the impact of survey delivery on the representation of these demographics was not significant. This means that these differences cannot be accredited to the survey delivery. However, the results do not suffice to reject the possibility of response errors. Higher response rates, especially when they do not approximate complete response like in this study, merely reduce the likelihood of unrepresentative results (Schouten et al., 2009). As it is previously shown that men, ethnic minorities and less highly educated individuals are generally underrepresented in different types of web-surveys due to their non-response (Bandilla et al., 2014), this tendency could still cause a misrepresentation in the twofold survey delivery. Nevertheless, the increased response bears relevance for the representativeness of the demographic data, since datasets with lower proportions of missing data provide a more solid basis for statistical methods to account for non-response and to yield reliable results (Schouten et al., 2009).

Furthermore, the results indicated of that the twofold survey delivery estimated a different association between parental education and completion and represented, on average, higher grades for learners from developing countries than the email-based survey. This indicates that response errors might complicate the validity of evidence on the association between the demographic background of MOOC learners and their learning outcomes (e.g., Van de Oudeweetering & Agirdag, 2017). This serves to suggest that representativeness deserves closer investigation in analyses with MOOC data. In general, the twofold survey appeared to reach a different part of the MOOC learning population, as it reached learners with lower average grades than the email-based survey. As previous findings indicated that

(30)

response was especially low among those with lower grades (Gates et al., 2014), the twofold survey delivery might be a solution to reach these learners.

5.2. Limitations and suggestions for future research

Although the study has provided relevant insights into the importance of survey delivery for the outcomes of MOOC studies, there were some limitations that need to be considered when interpreting the results. Furthermore, these limitations may help to inform about relevant directions for future research.

First, the results are not sufficient to conclude or dismiss a response error in either one of the survey deliveries, due to the considerable degree of non-response. However, some degree of non-response can be expected when keeping surveys voluntary. Although this seems to complicate knowledge on potential response errors, there are still research opportunities to advance knowledge on the representativeness of MOOC surveys. For example, respondents and non-respondents could be compared on background characteristics like their profile settings and activity patterns in order to make inferences about representativeness. Statistical methods like response-propensity models could further help to estimate and account for response errors (Kizilcec, 2014). Especially since increasing response can improve the reliability of these techniques (Schouten et al., 2009), the outcomes for different survey deliveries may be compared to check for the value of survey delivery in these procedures. However, approval of an Ethics committee on the use of non-respondents’ background data, as well as the adoption of a passive informed consent, is needed in order to conduct such analyses. Future research is therefore encouraged to anticipate these ethical concerns.

Furthermore, there were some technical difficulties that might have had consequences for the results. First, it was impossible to send the email-based survey to only one specific branch

(31)

(cf. experimental condition). Therefore, the alternative embedded survey had to be combined with the email-based survey. Although the findings indicate that mainly the embedded survey yielded higher response rates, it could be that its integration with the email-based survey held specific effects. For example, it could be that email invitations reminded the learners in experimental condition to respond to the embedded survey. Hence, as soon as the technical functionalities in MOOC platforms allow sending emails to the specific branches, it will be relevant to examine the effect of the embedded survey delivery separately. A second technical issue was that a substantial proportion of learners in the selected MOOC sessions received no or only one email. In order to ensure that all learners in the sample experienced the same research conditions, a substantial share of the sample had to be eliminated. In turn, this could have reduced the power of the outcomes. Finally, the study initially intended to include employment status as a learner demographic. However, due to technical difficulties, the data did not reveal which response was selected and employment status as a demographic indicator could not be included. Hence, future research is recommended to anticipate these technical limitations and look for solutions to enhance the precision, power and comprehensiveness of the results.

A final limitation that is anticipated concerns the scope of the study, as only MOOCs facilitated by the University of Amsterdam on the platform Coursera were included. As a consequence, the findings do not represent conclusions that can be generalized to other MOOCs and other platforms. Still, the inclusion of courses in different content areas and the consideration of between-course differences enabled the study to demonstrate potential course-level effects. For example, there were differences between courses in their representation of learners from developing countries, from different age groups, from both sexes and with different educational backgrounds. Moreover, learners in the different courses showed to vary in their average grades. Hence, the findings appeal to further research on the

(32)

impact of course-level characteristics on the enrollment and learning outcomes of learners with specific demographic characteristics.

5.3. Conclusion

This study provides a unique contribution to the literature on MOOCs and on web-surveys in general. This study was innovative in nature, as no previous research has examined the impact of alternative survey deliveries on response and response errors in demographic MOOC surveys. A specific quality of this study was its research design. Using a randomized controlled experiment rather than a cross-sectional design helped to improve the validity of the results by ruling out history and maturation effects. Furthermore, the study was able to detect and account for differences between MOOCs in different subject domains.

By focusing on survey deliveries that might make response more inviting and convenient, this study provided insight on a new strategy to raise response and increase likelihood of representative findings in MOOC research without violating the ethical consideration of voluntary research participation. Hence, the study could inform research in MOOCs and other online contexts to replace email-based surveys with other deliveries that are less time-consuming, more visible and more persuasive. More importantly, this study breaks new ground by raising awareness on possibilities to increase response rates and to examine representativeness in MOOC studies. By these means, the study provides well-informed reasons and strategies to scrutinize the validity of research in MOOCs and to improve its power to make general inferences on their learner population. This may help to gain more accurate knowledge on the potential of MOOCs to reach learners in disadvantaged positions and who fall out of the traditional higher education system. In this way, doubts or expectations about MOOCs as tools to enhance social equality in education and to improve social mobility may receive more accurate answers.

(33)

References

Annear, M. J., Toye, C. M., Eccleston, C. E., McInerney, F. J., Elliott, K. E. J., Tranter, B. K., Hartley, T., & Robinson, A. L. (2015). Dementia Knowledge Assessment Scale: Development and Preliminary Psychometric Properties. Journal of the American

Geriatrics Society, 63(11), 2375–2381.

Bandilla, W., Couper, M. P., & Kaczmirek, L. (2014). The Effectiveness of Mailed Invitations for Web Surveys and the Representativeness of Mixed-Mode versus Internet-only Samples. Survey Practice, 7(4), 1–9.

Banerjee, A. V., & Duflo, E. (2014). (Dis) Organization and success in an economics MOOC. The American Economic Review, 104(5), 514–518.

Belanger, Y., & Thornton, J. (2013). Bioelectricity: A quantitative approach Duke University’s first MOOC. Retrieved October 17, 2015, from

http://dukespace.lib.duke.edu/dspace/bitstream/handle/10161/6216/Duke_Bioelectricit y_MOOC_Fall2012.pdf

Christensen, G., Steinmetz, A., Alcorn, B., Bennett, A., Woods, D., & Emanuel, E. J. (2013). The MOOC Phenomenon: Who Takes Massive Open Online Courses and Why ? Retrieved October 17, 2016, from

http://papers.ssrn.com/sol3/papers.cfm?abstract_id=2350964

Cisel, M., Bachelet, R., & Bruillard, E. (2014). Peer assessment in the first French MOOC: Analyzing assessors' behavior. In Stamper, J., Pardos, Z., Mavrikis, M., McLaren, B.M. (Eds.), Proceedings of the 7th International Conference on

Educational Data Mining (pp. 403–404). London, UK: EDM.

Colas, J. F., Sloep, P. B., & Garreta-Domingo, M. (2016). The Effect of Multilingual

Facilitation on Active Participation in MOOCs. The International Review of Research

(34)

Couper, M. P., Kapteyn, A., Schonlau, M., & Winter, J. (2007). Noncoverage and nonresponse in an Internet survey. Social Science Research, 36(1), 131–148.

Dillahunt, T. R., Wang, B. Z., & Teasley, S. (2014). Democratizing higher education: Exploring MOOC use among those who cannot afford a formal education.

International Review of Research in Open and Distributed Learning, 15(5), 177–196. Fan, W., & Yan, Z. (2010). Factors affecting response rates of the web survey: A systematic

review. Computers in Human Behavior, 26(2), 132–139.

Graham, J. W., Olchowski, A. E., & Gilreath, T. D. (2007). How many imputations are really needed? Some practical clarifications of multiple imputation theory. Prevention

Science, 8(3), 206–213.

Grandcolas, U., Rettie, R., & Marusenko, K. (2003). Web survey bias: sample or mode effect? Journal of marketing management, 19(5-6), 541–561.

Greene, J. A., Oswald, C. A., & Pomerantz, J. (2015). Predictors of retention and achievement in a massive open online course. American Educational Research Journal, 52(5), 925– 955.

Galesic, M., & Bosnjak, M. (2009). Effects of questionnaire length on participation and indicators of response quality in a web survey. Public opinion quarterly, 73(2), 349– 360.

Gates, K., Wilkins, D., Conlon, S., Mossing, S., & Eftink, M. (2014). Maximizing the value of student ratings through data mining. In A. Peña-Ayala (Ed.), Educational Data Mining (pp. 379–410). Cham: Springer International Publishing.

Groves, R. M., Cialdini, R. B., & Couper, M. P. (1992). Understanding the decision to participate in a survey. Public Opinion Quarterly, 56(4), 475–495.

(35)

Han, V., Albaum, G., Wiley, J. B., & Thirkell, P. (2009). Applying theory to structure respondents' stated motivations for participating in web surveys. Qualitative Market Research: An International Journal, 12(4), 428–442.

Ho, A. D., Chuang, I., Reich, J., Coleman, C. A., Whitehill, J., Northcutt, C. G., Williams, J.J., Hansen, J., Lopez, G. & Petersen, R. (2015). Harvardx and MITx: Two years of open online courses fall 2012-summer 2014. Retrieved May 28, 2016, from http://papers.ssrn.com/sol3/papers.cfm?abstract_id=2586847

Kay, J., Reimann, P., Diebold, E., & Kummerfeld, B. (2013). MOOCs: So Many Learners, So Much Potential. IEEE Intelligent Systems, 28(3), 70–77.

Kizilcec, R. F. (2014). Reducing nonresponse bias with survey reweighting: Applications for online learning researchers. In M. Sahami, A. Fox, M.A. Hearst, & M.T.H. Chi (Eds.),

Proceedings of the first ACM conference on Learning@ scale conference (pp. 143– 144), Atlanta, GA: ACM

Kizilcec, R. F., Piech, C., & Schneider, E. (2013). Deconstructing disengagement: analyzing learner subpopulations in massive open online courses. In D. Suthers, K., Verbert, E. Duval, & X. Ochoa (Eds.), Proceedings of the third International Conference on Learning Analytics and Knowledge (pp. 170–179). Leuven, Belgium: ACM.

Kizilcec, R. F., Saltarelli, A. J., Reich, J., & Cohen, G. L. (2017). Closing global achievement gaps in MOOCs. Science, 355(6322), 251–252.

Laguilles, J. S., Williams, E. A., & Saunders, D. B. (2011). Can lottery incentives boost web survey response rates? Findings from four experiments. Research in Higher Education, 52(5), 537–553.

Porter, S. R. (2004). Raising response rates: what works? New directions for institutional research, 2004(121), 5–21.

(36)

Porter, S. R., & Umbach, P. D. (2006). Student survey response rates across institutions: Why do they vary? Research in Higher Education, 47(2), 229–247.

Reips, U. D. (2000). Context effects in Web-surveys. In B. E., Batinic, U. D. E. Reips, & M. E. Bosnjak (Eds.), Online social sciences (pp. 95–104). Göttingen: Hogrefe & Huber Publishers.

Rizzardini, R. H., Gütl, C., Chang, V., & Morales, M. (2014). MOOC in Latin America: Implementation and lessons learned. In L., Uden, Y. H. Tao, H.C. Yang, & I. H. Ting (Eds.), The 2nd International Workshop on Learning Technology for Education in Cloud (pp. 147–158). Amsterdam: Springer Netherlands.

Ross, J., Sinclair, C., Knox, J., & Macleod, H. (2014). Teacher experiences and academic identity: The missing components of MOOC pedagogy. Journal of Online Learning

and Teaching, 10(1), 57–69.

Sax, L. J., Gilmartin, S. K., & Bryant, A. N. (2003). Assessing response rates and nonresponse bias in web and paper surveys. Research in Higher Education, 44(4), 409–432.

Schouten, B., Cobben, F., & Bethlehem, J. (2009). Indicators for the representativeness of survey response. Survey Methodology, 35(1), 101–113.

Shrader, S., Wu, M., Owens-Nicholson, D., & Santa Ana, K. (2016). Massive open online courses (MOOCs): Participant activity, demographics, and satisfaction. Online

Learning, 20(2), 215–233.

Shah, D. (2015). MOOCs in 2015: Breaking down the numbers. Retrieved November 10, 2016, from https://www.edsurge.com/news/2015-12-28-moocs-in-2015-breaking-down-the-numbers

Snijders, T.A.B., & Bosker, R.J. (2012). Multilevel analysis. An introduction to basic and advanced multilevel modeling (2nd Edition). London: Sage Publications Ltd.

(37)

Stevens, J. P. (2009). Applied multivariate statistics for the social sciences. London: Routledge.

United Nations (1999). Standard Country or Area Codes for Statistical Use (M49). Retrieved March 23, 2017, from https://unstats.un.org/unsd/methodology/m49/

Van Buuren, S., & Groothuis-Oudshoorn, K. (2011). Mice: Multivariate imputation by chained equations in R. Journal of statistical software, 45(3), 1–17.

VanVoorhis, C. W., & Morgan, B. L. (2007). Understanding power and rules of thumb for determining sample sizes. Tutorials in Quantitative Methods for Psychology, 3(2), 43– 50.

Van de Oudeweetering, K., & Agirdag, O. (2017). MOOCs as accelerators of social mobility? A systematic review. Journal of Education, Technology and Society. Forthcoming. Ziegenfuss, D. H. (2016). Closing the Loop: Building Synergy for Learning through

Professional Development MOOC about Flipped Teaching. Current Issues in

Referenties

GERELATEERDE DOCUMENTEN

Moreover, I have used a LN transformation for the different types of compensation and total assets (firm size) to reduce the impact of a skewed distribution..

For example, the effect of parents' educatIOn on the educational attainment of Hungarian men was large in the older cohorts (around 1940) but is almost absent in the most recent

This is the certified public account­ ants certificate (C.P.A.) which is conferred by 54 State Boards of Accountancy - one in each of the fifty states and four others covering

Larger expectation gaps concern the fairness of the complaints procedure: 39 % of the public say that com- plaints are dealt with through fair procedures and 37 % say that

Our research aimed to give young patients an opportunity to speak up for themselves in the development process of two new patient experience surveys on rehabilitation care

More specifically, the more the consumers perceive the brand of the music festival to be differentiated, the more likely they are to have a greater purchase intention, a more

y-grOCIIO &lt;laarin sal slang om die poli ticl'c mondstuk ntn die g corganisccr- do \'alwnies tc word ttic... kels deur di e strate

- Product between policy index, log of the difference in population between donor country and receiving country and dummy for if a country was a colony of the United Kingdom. -