Skip to main content
Open AccessReview Article

Decrypting Log Data

A Meta-Analysis on General Online Activity and Learning Outcome Within Digital Learning Environments

Published Online:https://doi.org/10.1027/2151-2604/a000484

Abstract

Abstract. Analyzing log data from digital learning environments provides information about online learning. However, it remains unclear how this information can be transferred to psychologically meaningful variables or how it is linked to learning outcomes. The present study summarizes findings on correlations between general online activity and learning outcomes in university settings. The course format, instructions to engage in online discussions, requirements, operationalization of general online activity, and publication year are considered moderators. A multi-source search provided 41 studies (N = 28,986) reporting 69 independent samples and 104 effect sizes. The three-level random-effects meta-analysis identified a pooled effect of r = .25 p = .003, 95% CI [.09, .41], indicating that students who are more active online have better grades. Despite high heterogeneity, Q(103) = 3,960.04, p < .001, moderator analyses showed no statistically significant effect. We discuss further potential influencing factors in online courses and highlight the potential of learning analytics.

Until recently, face-to-face teaching and on-site exams were considered the gold standard in formal higher education. Universities face increased demand as the number of students constantly grows (Araka et al., 2020), and because of the COVID-19 pandemic, the need for online learning in higher education increased drastically (Ali, 2020). However, only little is known about how students use online classes and how their learning behavior is linked to learning outcomes. Online learning environments, so-called Learning Management Systems (LMS), make it easier to obtain information about students’ learning behavior by analyzing automatically tracked log data from students’ interactions with the LMS (Gašević et al., 2016). Although interest in learning analytics constantly grows (Dawson et al., 2014) and uses of trace data are increasing (Winne, 2020), the potentials of using process generated data for purposes of summative and formative assessments is yet an emerging research field in which common concerns about the potential limitations are addressed. Concerns relate to the weak relation between overt learning behavior operationalized via log data with complex learning behavior. Henrie and colleagues (2018) describe them as strongly simplified proxies for complex learning behavior, which ultimately questions their usefulness for the evaluation of online classes. Consequently, there is an ongoing debate whether log data are a valid predictor for learning outcomes (e.g., Agudo-Peregrina et al., 2014; Campbell, 2007), or whether they fail to predict learning outcomes – either because it is difficult linking log data to learning behavior, or because courses with substantial online elements are too heterogeneous to draw a general conclusion (e.g., Gašević et al., 2016; Macfadyen & Dawson, 2010). Although added value of using fine-grained and event-related process data for certain unobtrusive assessment purposes has been demonstrated (e.g., measuring intra-individual change; Barthakur et al., 2021), the use of log data for platform-, domain- and demand independent assessment of successful digital learning activities is still a matter of debate. Since systematic reviews about the value of broad log indicators are missing, the present meta-analysis summarizes findings on the relationship between these online activities derived from log data and learning outcomes within LMS. We focus on two broad log data indicators of general online activity (i.e., total login time and login frequency), which can be classified as access-related log events (Kroehne & Goldhammer, 2018) and are commonly used as measures linked to students’ achievement (see You, 2016 for a detailed discussion of this issue). Accordingly, in the case of learning outcomes, we focus on indicators of learning success (i.e., course grade or course score).

General Online Activity and Learning Outcomes

One major advantage of online learning for educational research is the availability of a vast amount of information that can be derived automatically and unobtrusively. For example, courses offered via LMS allow collecting an enormous amount of log data about interactions with the platform (Campbell, 2007). However, the primary use of explorative approaches for analyzing log data results in a lack of theoretical grounding (Winne, 2020). Besides the number of studies comprising data-driven approaches for the decision which log data to examine, several researchers have considered pedagogical theories (see Tempelaar et al., 2015, for a review). For example, general online activities, such as total login time or login frequency, are considered indicators for learning engagement (Beer et al., 2010). Furthermore, according to Carroll’s (1963)model of school learning, time spent on learning is one of the crucial factors for students’ performance. Thus, time spent learning online might also be a crucial factor for web-based achievement (Jo et al., 2015). Besides, login frequency is associated with Engeström’s (1987)activity theory, which states that mere activity produces meaningful learning, leading to higher learning outcomes (Kupczynski et al., 2011). Finally, total login time and login frequency are considered proxies for time management strategies, as they are indicators for sufficient time investment and, thus, key factors for performance (Jo et al., 2015). However, there is a debate on the type of log data suitable to measure learning behavior (Agudo-Peregrina et al., 2014). Critics suggest focusing on quality, not the quantity of online learning behavior (You, 2016): As active participation is crucial for success, indicators that do not distinguish between active and passive engagement are problematic (Ransdell & Gaillard-Kenney, 2009). Overall, these contradictory assumptions on the usefulness of broad log data indicators go along with inconsistent findings on the association between those indicators of general online activity and learning outcomes: some studies reported no association (Broadbent, 2016), negative correlations (e.g., Ransdell & Gaillard-Kenney, 2009; Strang, 2016), or positive correlations (e.g., Liu & Feng, 2011; McCuaig & Baldwin, 2012; Saqr et al., 2017). Other studies that examined various online courses simultaneously obtained mixed results across different courses (e.g., Conijn et al., 2017; Gašević et al., 2016), indicating that online courses might be too heterogeneous to draw a general conclusion about the link between general online activity and learning outcomes.

The Present Study

Our aim was to systematically review findings on the relationship between two log data indicators of general online activity and learning outcomes within LMS. To guarantee a minimum level of comparability between classes, we focused on online university courses because they are more structured than informal online courses (Song & Bonk, 2016) and because they are usually graded, offering a measure of learning outcomes. Regarding general online activity, we focused on total login time and login frequency to assess the applicability of these broad measures derived from log data as a proxy of online learning behavior and examine how they are linked to educational outcomes (i.e., course grade or course score). Moreover, we examined the impact of several moderators to explain the inconsistent findings reported in previous literature, since in the course of recent technological developments (Palvia et al., 2018), teaching tools became more sophisticated and online courses became more diverse.

First, we use the term “online course” to describe all courses that include substantial online elements, that is, courses that are taught exclusively online (fully online courses), and courses that combine online and face-to-face delivered content (blended courses; Allen & Seaman, 2014). Compared to fully online courses, blended courses offer more structure through regular face-to-face sessions (Means et al., 2013), as well as the opportunity to easily get in touch with peers regularly (Broadbent, 2016). However, the varying parts of face-to-face teaching in blended courses cannot be tracked via log data (Mwalumbwe & Mtebe, 2017). Therefore, we expect a stronger relationship between general online activity and learning outcomes within fully online than within blended learning courses, as for the latter, substantial parts of learning might remain unreflected by log data (Hypothesis 1).

Second, online courses vary with respect to the emphasis put on the use of online discussion boards. An explicit instruction to use discussion boards implemented within the LMS might foster deeper learning while being online and lead to better achievement (Song et al., 2019). Although interactions in discussion forums are considered an essential part of learning (Uijl et al., 2017), the mere existence of a discussion board is not enough for promoting active participation within the LMS (Lee & Martin, 2017). Since active content engagement is crucial for students’ achievement (Ransdell & Gaillard-Kenney, 2009), we expect higher correlations between general online activity and learning outcomes for courses with instruction for discussion board usage (Hypothesis 2).

Third, online courses differ in their grading systems. For the present study, it is important to what extent online activities are explicitly incentivized. While some courses do not incentivize online participation at all, other courses either offer bonus points for regular online participation or even require online activities within the LMS (e.g., online group discussions, quizzes, or online assignments) as part of the final grade. These incentives encourage active online engagement (Tempelaar et al., 2019) and offer guidance for students to effectively use the tools and materials provided within the LMS. Therefore, we expect a higher correlation between general online activity and learning outcomes for courses that incentivize certain online activities through their grading systems (Hypothesis 3).

Fourth, we compare the operationalization of general online activity as total login time or as login frequency. Ever since the growing popularity of investigating log data, there has been the challenge of capturing log data that might be translated into psychologically meaningful variables (Seifert et al., 2018). As both operationalizations are theoretically reasoned with either Carroll’s (1963)model of school learning or Engeström’s (1987)activity theory, we want to explore if general online activity operationalized as total login time versus login frequency differs in their relationship with learning outcomes (Hypothesis 4).

Lastly, we considered publication year as a potential moderator. In the face of rapid technological advancements (Palvia et al., 2018), we expect changes in how LMS provides education. Through technological change, LMS tools become more advanced, and multiple types of learning tools can be implemented to foster students’ active engagement within the LMS (Kebritchi et al., 2017). Therefore, students ought to be enabled to benefit from a more interactive online learning experience, and online activity within recent studies might result in higher achievements than within older studies (Hypothesis 5).

Method

In accordance with common open science practices, we provide all additional materials (i.e., coding manual, syntax, data, PRISMA20-checklist, and supplemental figures and tables) online within the Open Science Framework (Center for Open Science, 2021).

Literature Search and Study Selection

We identified 33,724 potentially relevant studies from electronic searches in major scientific databases (PsycINFO, PsycArticles, PSYNDEX, and ERIC) and Google Scholar using the Boolean search term (online learning OR online course* OR web-based learning OR e-learning OR elearning OR learning management system* OR LMS OR learning analytics) AND (achievement OR performance OR outcome) in February 2021. We retrieved five further studies by calls for unpublished work (via mailing list of the German Psychological Society, ResearchGate, and Twitter). Additionally, we performed a “rolling snowball” search and identified 17 further studies by screening the reference lists of all eligible studies and by conducting forward citation tracking using Google Scholar. Finally, we contacted 19 corresponding authors of studies not reporting bivariate correlations and received them for three studies. We included all published or unpublished types of studies. See Figure S1 for the detailed literature search process, including specifications of all sources that were searched. Subsequently, these studies were included in the analysis depending on the following inclusion criteria: (a) The study investigated a fully online or blended course in an institutional setting; (b) General online activity was measured using log data and operationalized as total login time or login frequency (i.e., number of single logins or number of days with at least one login); (c) Learning outcome was measured as course grade or course score; (d) The study consisted of a sample comprising university students; (e) The study was published between 1969 (year of the first connection of the Internet) and 2021, and (f) was written in English or German; (g) The study reported at least one correlation between general online activity and learning outcome or appropriate statistics that could be transformed into correlations. Studies were excluded if: (h) General online activity was measured as a self-report as we focused on the usefulness of log data indicators of general online activity, or (i) measured as the duration or frequency of single activities in the LMS (e.g., time spent on quizzes, number of forum postings) because these types of log data fall within different categories (such as response-related or process-related log data; see Kroehne & Goldhammer, 2018), and (j) the study had a commercial e-learning course as setting as we focused on the specific context of higher education. After applying these criteria, 41 primary studies remained (see Table S1). Study selection followed a two-stage process. First, two researchers reviewed titles and abstracts of the first 50 records and discussed disagreements about eligibility until consensus was reached. Then, one researcher screened all titles and abstracts of all studies retrieved. In cases in which eligibility was unclear, the study was considered for the second stage in the form of a full-text review. A sample of full-text studies (~15%; 36/241) was independently screened by two researchers. The remaining full-text studies were screened by one researcher. Finally, the second researcher independently reviewed all included studies and those with uncertain eligibility. Again, disagreements about eligibility were resolved through discussion.

Coding Process

We developed a standardized coding manual and data extraction sheet for the data collection process (see Center for Open Science, 2021 for detailed information). The coding manual comprised eligibility criteria, guidelines for selecting effect sizes and coding, and definitions of all outcomes and other variables for which data were sought (i.e., name and description of the respective variables, guidelines regarding the format of coding, and coding examples). For each study, we extracted all relevant effect sizes for the association between general online activity and learning outcomes. Moreover, we collected data on study and sample and online course characteristics covering especially moderator variables (i.e., course format, emphasis of discussion, course activities as part of grading, operationalization of general online activity, and publication year) and general study information.

All studies were coded twice using the coding manual and data extraction sheets by two independent raters. To evaluate the coding process, Cohen’s (1960) κ for categorical variables and intraclass coefficients (ICC; Shrout & Fleiss, 1979) for continuous variables were calculated for the focal variables. Interrater reliability for the effect size was ICC = .89, 95% CI [.84, .92] and for the sample size and publication year ICC = 1, 95% CI [1, 1]. The Cohen’s κ for the remaining categorical variables was .91, overall indicating strong to excellent intercoder agreement (LeBreton & Senter, 2008). All discrepancies were solved upon discussion by comparing extracted data.

Statistical Analyses

Effect Size

We used Pearson product-moment correlation as an effect size measure. Because transforming standardized weights from multiple linear regression analyses into correlation coefficients is problematic (Aloe, 2015), authors from studies reporting only regression weights were contacted to obtain correlations. If no correlation was available, the study was excluded from the analyses (k = 6). To standardize the direction of effects, we conversed effect sizes in cases where learning outcomes were conceptualized as smaller numbers indicating better achievement.

Meta-Analytic Model

We pooled effect sizes using a random-effects model with a restricted maximum likelihood estimator (Viechtbauer, 2010). A three-level meta-analysis was conducted to account for dependent effect sizes (Cheung, 2014) because some studies reported more than one effect size (e.g., provided correlations for both operationalizations of general online activity for a given sample). Dependencies between effect sizes derived from the same sample are acknowledged by decomposing the total random variance into two variance components: one reflecting the heterogeneity of effects within samples, and the other indicating heterogeneity of effect sizes between samples (see Gnambs & Appel, 2018 for a detailed description). We calculated I2 statistics to quantify heterogeneity in observed effect sizes (Higgins et al., 2003). Considering I2 is not an absolute measure of heterogeneity (Borenstein et al., 2017), we additionally report the Q-statistics. Since using sample size weights performs best for estimating the random-effects variance component in meta-analytic models with correlations as effect size measures (Brannick et al., 2011), we used this weighting procedure to account for sampling error. We reported our findings focusing on the size of the effect and its confidence and prediction interval. To visualize our meta-analysis, we used a forest plot (Viechtbauer, 2010). Lastly, we conducted subgroup and meta-regression analyses to examine moderating effects on the pooled effect size (Harrer et al., 2019), given the diversity of online courses being investigated. Therefore – apart from publication year –, we categorized the included studies along with dichotomous moderators: fully online vs. blended course format, instructed vs. not instructed discussion board usage, graded activities vs. no requirements, and total login time vs. login frequency as general online activity.

Sensitivity Analyses

First, we used the studentized deleted residuals (Viechtbauer & Cheung, 2010) to identify extreme correlations. Additionally, we conducted sensitivity analyses that removed the identified outliers from the analyses to examine the impact of these outliers. Moreover, the robustness of the presented meta-analysis was investigated by removing two particular studies that differed in their conceptualization (i.e., Lauría et al., 2012: correlation comprised data from an entire university; and Mödritscher et al., 2013: log data from only two weeks before the examination) from the meta-analytic database and comparing the pooled effect to the pooled effect from the full database.

Publication Bias

The presence of potential publication bias was investigated in two ways: First, we performed a meta-regression with the publication type as a moderator. Effect sizes from peer-reviewed sources were compared to effect sizes from other sources (i.e., theses or conference papers). A statistically significant difference between effect sizes extracted from both sources could result from a distortion in the peer-reviewed research literature due to systematic suppression of (e.g., nonsignificant) effects. Second, we conducted a modified regression test for asymmetry by including a precision measure (i.e., 1/n) as a moderator in the meta-analytic model to account for dependent effect sizes.

Statistical Software

All analyses were conducted using R version 4.0.4 (R Core Team, 2021). Meta-analytic models were estimated with the metafor package version 2.4-0 (Viechtbauer, 2010).

Results

Descriptive Statistics

The meta-analysis is based on 41 studies published between 1997 and 2021, predominantly as peer-reviewed articles (73%). The remaining studies appeared as theses (5%) or conference papers (22%). The database covered 69 independent samples that provided 104 effect sizes, with each sample comprising between 1 and 3 effect sizes. Overall, the meta-analysis included scores from 28,986 students (range of samples’ ns: 11–11,195, Mdn = 122). The mean age was 22.21 years, and 53.71% of the students were female, however, only 11 studies reported information on age, and 20 studies information on gender. The duration of the courses varied between 6 and 19 weeks (Mdn = 12 weeks), mainly covering one academic semester. Moreover, courses varied with respect to their format (24% fully online, 72% blended, 1% not reported separately, or 3% missing), emphasis of discussion (18% instructed use of discussion boards, 69% discussion boards available within the LMS without further instructions, 10% not mentioned, or 3% missing), and requirements (45% online activities within the LMS as part of grading, 54% none, or 1% missing). In 44% of the cases, general online activity was operationalized as total login time and 56% as login frequency (number of single logins or number of days with at least one login).

Overall Pooled Correlation

In total, the three-level random-effects meta-analysis identified a pooled correlation of r = .25 p = .003, 95% CI [.09, .41], indicating that students who are more active online also have a better learning outcome (Figure 1). The result of the pooled correlation was robust and replicated in the separate moderator analyses (Table 1).

Figure 1 Forest plot. Note. Effect sizes are ordered by increasing magnitude. Larger symbols illustrate larger sample sizes.
Table 1 Meta-analysis on general online activity and learning outcome and separate moderator analyses

Overall, these findings indicate a small but statistically significant positive association between general online activity and learning outcomes. Yet, the high random variance resulted in an exceedingly large prediction interval around the pooled effect, 80% PI [−.10, .59]. Hence, we conducted sensitivity analyses to examine the impact of certain studies on the prediction interval. Further, the studies showed higher between-cluster heterogeneity (I2 = .92; see also Figure 1 for an illustration of the variability between samples) compared to within-cluster-heterogeneity (I2 = .05), indicating pronounced unaccounted differences between samples that might be explained by moderator analyses, but negligible variability within samples (Higgins et al., 2003).

Moderator Analyses

We conducted meta-regression analyses to examine the effects of course format, emphasis of discussion, requirements, operationalization of general online activity, and publication year on the pooled effect (Table 2). On effect size level, correlations between moderators ranged from −.54 to .46, indicating negligible multicollinearity. None of the moderators was statistically significant. This result remained the same even when each moderator was examined separately (Table 1). Overall, moderator analyses showed no effect, indicating that our data do not provide evidence in favor of a moderating effect of course format, emphasis of discussion, requirements, operationalization of general online activity, or publication year on the relationship between general online activity and learning outcome.

Table 2 Moderator analysis including all five moderator variables simultaneously

Sensitivity Analyses

We performed sensitivity analyses to examine the robustness of our findings. In a first step, the robustness of the presented results was investigated by removing nine extreme correlations (i.e., outliers with z > 1.96; Viechtbauer & Cheung, 2010; Figure S2) from the database to compare this pooled effect to the original pooled effect. After eliminating these effects from the database, the pooled effect was r = .24, p < .001. The 80% PI decreased from [−.10, .59] to [.04, .43], indicating a reduced random variance. However, the outliers did not distort the pooled effect. Similar patterns appeared for all subgroup analyses (see Table S2). Overall, the outlier analyses provided evidence for the robustness of the correlation between general online activity and learning outcomes. In a second step, sensitivity analyses with respect to two studies that differ in their conceptualizations from the other included studies resulted in negligible differences: r = .29, p < .001, 80% PI [−.05, .62] (Lauría et al., 2012), and r = .24, p = .011, 80% PI [−.11, .60] (Mödritscher et al., 2013), also indicating the robustness of the present meta-analysis (see Table S3).

Publication Bias

First, the meta-regression analysis that we conducted to examine publication bias indicated no statistically significant difference between effect sizes extracted from peer-reviewed versus other sources (γ = −0.06, SE = 0.13, p = .624). Second, the modified regression test for asymmetry (γ = 7.93, SE = 8.57, p = .355) revealed no statistically significant effect for measurement precision. Overall, we did not find evidence of publication bias.

Discussion

Online courses are more important than ever (Ali, 2020), and they provide the possibility to conveniently and unobtrusively record log data (Dawson et al., 2014). Yet, it is unclear how log data contribute to explaining the linkage of learning behavior to academic achievement. Although several studies have examined the association between general online activities and learning outcomes, their findings are ambiguous (e.g., Broadbent, 2016; Campbell, 2007; Gašević et al., 2016). Hence, we provided a systematic review of existing findings and investigated several potential moderators to explain ambiguity in previous literature: In a first step, we identified a small – yet statistically significant – pooled correlation of r = .25 between general online activity and learning outcome. This finding indicates that students who are online for a longer time (or more often) within the LMS also tend to have better course grades. This effect might seem small at first, but it remained robust across sensitivity analyses even though we used very broad indicators of general online activity. Additionally, academic success in itself is extremely complex and, therefore, difficult to predict (see Alyahyan & Düştegör, 2020, for a review). Comparing our results to a meta-analysis examining multiple psychological correlates of university students’ academic performance (Richardson et al., 2012), log data indicators perform better in predicting academic success as compared to demographics (i.e., gender, age, and socioeconomic status) and personality traits (except for conscientiousness), but perform slightly worse than prior academic performance and academic self-efficacy. The strongest correlate of all 50 measures was performance self-efficacy (r = .59). Against this background, the present analysis demonstrates the potential of log data, given that even two broad log data indicators of online learning behavior are associated with the learning outcome. However, the meta-analytic model revealed high heterogeneity between studies that could not be explained by moderator analyses. Therefore, we discuss reasons why our moderator variables might have failed to explain high heterogeneity and other possible sources of variance.

Limitations of the Included Moderators

First, our potential moderator variables were restricted to broad course characteristics, which can be illustrated by the variable course format. The dichotomous classification of blended versus fully online format might be too coarse as there exist flowing transitions depending on the portion of content delivered online (Allen & Seaman, 2014) and the share of online elements in a certain course might be better depicted as a continuous variable rather than a dichotomous one. However, most studies on online courses only provide very superficial characteristics. Given the fact that faculties struggle with the transition to online teaching (Kebritchi et al., 2017), more evidence is needed so practitioners who design online courses are able to make informed choices to improve quality of online learning.

Second, there is a lack of information on contextual variables (e.g., instructional design, interactive tools, or synchronicity) reported in primary studies. Varying contexts and tools affect the learning process by providing different learning opportunities which are decisive for improved learning (Lust et al., 2012). The consideration of contextual factors might help explain ambiguous findings in the current literature (Gašević et al., 2016) as they enable more comprehensive moderator analyses. Our moderator analyses were limited to basic information about how the LMS was implemented. In the following, we discuss which aspects might help explain differences between settings.

Potential Other Moderators

An overall structure offered by online courses might help reduce individual differences in online learning behavior, as it provides guidance for students to engage in the most beneficial activities at a certain point of course processing (Winne, 2004). On the individual level, instructors can help to reduce existing heterogeneity within the association between general online activity and learning outcome, as not all learners seem to benefit equally from learning opportunities (Lust et al., 2012). Learners need to be instructed how to use LMS (Kebritchi et al., 2017) to exploit the full potential of online courses. Other forms of structure are, for example, shares of synchronous methods and applications in online courses (e.g., teaching sessions, collaborative learning, or support and monitoring by a tutor; see Kinshuk & Chen, 2006), which provide a structured schedule for learning behavior, or any form of online assessment across the course duration to monitor students’ learning progress or to provide personalized feedback (Knight, 2020), or to encourage students’ engagement (Tempelaar et al., 2019). In the present literature, systematic information on the extent of the structuredness of online courses is unfortunately limited. Future research might document the effects of these course characteristics and their impact on learning behavior.

Another aspect of online course design comprises the incentives that are used to ensure students’ participation. Apart from the extent to which participation is included in course grading, only little is known about how instructors use incentives for constant participation throughout the course. One example for these incentives is gamification – the implementation of game-design principles and elements in non-game environments (Deterding et al., 2011) – which can promote motivation (see Mora et al., 2017 for a review) for example by providing visualized immediate feedback to the learner on goal completion or students’ learning progress compared to other students. Gamification for educational purposes can be associated with increased activity (e.g., Hamari, 2017; Huang & Hew, 2015) or general engagement in online programs (Looyestyn et al., 2017). But it remains unclear if a game-based induced increase in online activity automatically leads to improvement in learning. If practitioners systematically provide online courses with and without different types of gamification, future research could examine differences in the online learning activity and its impact on learning outcomes.

Finally, our meta-analysis was based on log data indicators that specify the extent of total login time or login frequency within LMS over an entire academic semester. Differences in the distribution of online activity across the course duration could not be considered. As distributed learning is a more efficient learning strategy than cramming before examinations for an equal amount of time (Dunn et al., 2013), future research should address the mere amount of activity and the distribution of total login time or logins in a more fine-grained way. Additionally, students’ diversity and consistency of online activities might provide substantial insights into how students’ activity affects learning (Lust et al., 2012).

Log Data in Educational Research

The present meta-analysis can be seen as a starting point of how log data can be used to contribute to our understanding of complex variables like academic achievement by linking course outcomes to broad log data indicators of online learning behavior. However, given the increase of online education in higher education and the recent technological development (Kebritchi et al., 2017) there undoubtedly exist more fine-grained data in research than overall participation measures comprising platform usage. One big advantage of the use of log data is that large amounts of data are easily and immediately accessible so that learning analytics can draw on detailed and extensive log data about learners’ studying activities within modern software tools (Winne, 2010). Although log data are a more accurate reflection of the quantity of media use than self-reports (Parry et al., 2021), researchers’ degrees of freedom in data tracking, collection and analysis persist and thereby limiting the objectiveness of log data indicators (Avella et al., 2016). Moreover, the biggest issue is the connection with existing educational theories and the resulting necessity to consider the reliability of log data as well as its role in claims about validity based on this kind of data (Winne, 2020).

But how can learning analytics meet the vision to help improving learning and teaching by using generated data as people engage in learning? On the student-centered level, learning analytics facilitate predictive modeling of course completion (Clow, 2013). Information on student’s previous educational experience and demographics, as well as data on online activity and formative and summative assessment, are combined and then used to develop interventions designed to improve retention and performance. This enables early intervention systems and personalized learning, as students receive real-time feedback on their learning progress (Arnold & Pistilli, 2012). Otherwise, instructors can take advantage of learning analytics by using them to identify areas in need of improvements regarding the curriculum as well as their own performance (Avella et al., 2016). Finally, the implementation of new tools or mechanisms can be checked (Song, 2018). Multiple types of learning tools enhance the learning experience (Hathaway, 2014), but it is not always the more, the better. Instructors have to consider which tool, design element, or multimedia will add to the learning process and which ones are distracting (Kebritchi et al., 2017).

Limitations of the Present Study and Implications for Future Research

Research on learning analytics is a promising approach for an advanced understanding of the learning process (Gašević et al., 2015). This meta-analysis provides an initial insight into the value of broad log data indicators of learning behavior. However, the present analyses come with limitations. Specifically, recent developments in meta-analytic methods suggest that it might be more adequate to model the hierarchical structure of the data by including the covariances of effect sizes derived from the same sample in the model, rather than using the default model, which assumes no covariances. On a more general stance, data dependency is an important issue in meta-analytic research that is often neglected (e.g., Rodgers & Pustejovsky, 2021). While these advanced models might be even better suited to model the data structures, many issues arise not from inappropriate modeling choices but rather from shortcomings in the primary studies. How should future research look like in order to contribute to a more advanced understanding of online learning?

While the number of studies using log data increases steadily, only a few of these studies transparently describe their methodologies for data collection and cleaning, utilized measures or analyses (Bergdahl et al., 2020). In general, learning analytics has to face challenges of heterogeneous data sources and the lack of unified vocabulary (Papamitsiou & Economides, 2014). Future meta-analyses could make use of quality assessments of primary studies, something that is already common for assessing the methodological quality of intervention effectiveness research (e.g., Valentine & Cooper, 2008). Scheffel and colleagues (2014) have proposed a framework of quality indicators for learning analytics earlier on. Future meta-analyses would benefit from a standardized procedure that allows taking the methodological quality of learning analytics studies into account for weighting procedures as well as the decision whether to include or exclude a primary study. However, learning analytics can benefit from a unified framework for the use of terms and definitions, operationalizations, and methodological procedures.

Another promising development to overcome central issues (i.e., data access and transparency) in meta-analyses is the open data movement (Gurevitch et al., 2018). As soon as researchers follow standards regarding an open scientific process, design standards would reduce unclearly reported methodologies, and data sharing standards would enable to directly generate effect sizes from open data (Nosek et al., 2015).

Moreover, open science practices facilitate multi-level analyses based on raw data in the form of meta-analysis of individual participant data (IPD; e.g., Riley et al., 2010). IPD meta-analyses are considered the gold standard as they prevent aggregation biases and enable to look at the impact of heterogeneity that originates from differences within studies (Kaufmann et al., 2016). Due to the COVID-19 pandemic, educational institutions were forced to shift teaching to online learning (De’ et al., 2020), potentially with an increase in the usage of interactive tools (such as video conferencing tools) that could facilitate online discussions or group work, which may also lead to an increase in the quality of online courses. Hopefully, these changes will be documented in forthcoming research on online courses. Future research might comprise large collaborations and centrally coordinated data collections within online courses to benefit from the incoming data due to the digital surge so that they might gain deeper insights to improve the quality of online learning.

Apart from that, this review focused on formal higher education. However, informal learning gains more attention in the field of educational research (Zheng et al., 2019), and therefore is a promising extension. Especially the change toward online learning promotes informal learning (i.e., learner-directed and independent learning outside of formal educational contexts; Song & Bonk, 2016). However, due to the absence of external assessment within informal learning (Callanan et al., 2011), studies with an informal context could not be included in the present meta-analysis. Even though formal and informal learning lead to gains in knowledge and skills (Cerasoli et al., 2018), it is difficult to combine them in meta-analyses as their outcomes are operationalized differently since for informal learning, educational success is traditionally defined as course completion (Henderikx et al., 2017). Accordingly, it seems worthwhile to examine whether the positive association between general online activities and learning outcomes can be transferred to an informal context.

Conclusion

In summary, we identified an association between broad log data indicators of general online activity and learning outcomes. Although several sensitivity analyses indicated the robustness of the present meta-analysis, the high heterogeneity between studies could not be explained by our moderator variables, which were limited to basic information on course implementation. We recommend for future research to form bigger collaborations and centrally collect data to conduct IPD meta-analyses to gain deeper insight into online learning. Learning analytics have the potential to provide more fine-grained data, but it is necessary to connect generated data to existing educational theories.

References

  • Agudo-Peregrina, Á. F., Iglesias-Pradas, S., Conde-González, M. Á., & Hernández-García, Á. (2014). Can we predict success from log data in VLEs? Classification of interactions for learning analytics and their relation with performance in VLE-supported F2F and online learning. Computers in Human Behavior, 31, 542–550. https://doi.org/10.1016/j.chb.2013.05.031 First citation in articleCrossrefGoogle Scholar

  • Ali, W. (2020). Online and remote learning in higher education institutes: A necessity in light of COVID-19 pandemic. Higher Education Studies, 10(3), 16–25. https://doi.org/10.5539/hes.v10n3p16 First citation in articleCrossrefGoogle Scholar

  • Allen, I. E., & Seaman, J. (2014). Grade change: Tracking online education in the United States (Annual Report No. 11). Babson Survey Research Group and Quahog Research Group, LLC. https://eric.ed.gov/?id=ED602449 First citation in articleGoogle Scholar

  • Aloe, A. M. (2015). Inaccuracy of regression results in replacing bivariate correlations. Research Synthesis Methods, 6(1), 21–27. https://doi.org/10.1002/jrsm.1126 First citation in articleCrossrefGoogle Scholar

  • Alyahyan, E., & Düştegör, D. (2020). Predicting academic success in higher education: Literature review and best practices. International Journal of Educational Technology in Higher Education, 17, Article 3. https://doi.org/10.1186/s41239-020-0177-7 First citation in articleCrossrefGoogle Scholar

  • Araka, E., Maina, E., Gitonga, R., & Oboko, R. (2020). Research trends in measurement and intervention tools for self-regulated learning for e-learning environments – Systematic review (2008–2018). Research and Practice in Technology Enhanced Learning, 15, Article 6. https://doi.org/10.1186/s41039-020-00129-5 First citation in articleCrossrefGoogle Scholar

  • Arnold, K. E., & Pistilli, M. D. (2012). Course signals at Purdue: Using learning analytics to increase student success. Proceedings of the 2nd International Conference on Learning Analytics and Knowledge – LAK ‘12, 267–270. https://doi.org/10.1145/2330601.2330666 First citation in articleCrossrefGoogle Scholar

  • Avella, J. T., Kebritchi, M., Nunn, S. G., & Kanai, T. (2016). Learning analytics methods, benefits, and challenges in higher education: A systematic literature review. Online Learning, 20(2), 13–29. https://eric.ed.gov/?id=EJ1105911 First citation in articleGoogle Scholar

  • Barthakur, A., Kovanovic, V., Joksimovic, S., Siemens, G., Richey, M., & Dawson, S. (2021). Assessing program-level learning strategies in MOOCs. Computers in Human Behavior, 117, Article 106674. https://doi.org/10.1016/j.chb.2020.106674 First citation in articleCrossrefGoogle Scholar

  • Beer, C., Clark, K., & Jones, D. (2010). Indicators of engagement. In C. H. SteelM. J. KeppellP. GerbicS. HousegoEds., Curriculum, technology & transformation for an unknown future. Proceedings ascilite Sydney 2010 (pp. 75–86). https://ascilite.org/conferences/sydney10/procs/Beer-full.pdf First citation in articleGoogle Scholar

  • Bergdahl, N., Nouri, J., Karunaratne, T., Afzaal, M., & Saqr, M. (2020). Learning analytics for blended learning: A systematic review of theory, methodology, and ethical considerations. International Journal of Learning Analytics and Artificial Intelligence for Education, 2(2), 46–79. https://doi.org/10.3991/ijai.v2i2.17887 First citation in articleGoogle Scholar

  • Borenstein, M., Higgins, J. P. T., Hedges, L. V., & Rothstein, H. R. (2017). Basics of meta-analysis: I2 is not an absolute measure of heterogeneity. Research Synthesis Methods, 8(1), 5–18. https://doi.org/10.1002/jrsm.1230 First citation in articleCrossrefGoogle Scholar

  • Brannick, M. T., Yang, L.-Q., & Cafri, G. (2011). Comparison of weights for meta-analysis of r and d under realistic conditions. Organizational Research Methods, 14(4), 587–607. https://doi.org/10.1177/1094428110368725 First citation in articleCrossrefGoogle Scholar

  • Broadbent, J. (2016). Academic success is about self-efficacy rather than frequency of use of the learning management system. Australasian Journal of Educational Technology, 32(4), 38–49. https://doi.org/10.14742/ajet.2634 First citation in articleGoogle Scholar

  • Callanan, M., Cervantes, C., & Loomis, M. (2011). Informal learning. Wiley Interdisciplinary Reviews: Cognitive Science, 2(6), 646–655. https://doi.org/10.1002/wcs.143 First citation in articleCrossrefGoogle Scholar

  • Campbell, J. P. (2007). Utilizing student data within the course management system to determine undergraduate student academic success: An exploratory study (Publication No. 3287222) [Doctoral dissertation]. Purdue University. ProQuest Dissertations & Theses. First citation in articleGoogle Scholar

  • Carroll, J. B. (1963). A model of school learning. Teachers College Record, 64(8), 723–733. https://psycnet.apa.org/record/1963-08222-001 First citation in articleCrossrefGoogle Scholar

  • Center for Open Science. (2021). Open Science Framework. https://osf.io/wy2px/ First citation in articleGoogle Scholar

  • Cerasoli, C. P., Alliger, G. M., Donsbach, J. S., Mathieu, J. E., Tannenbaum, S. I., & Orvis, K. A. (2018). Antecedents and outcomes of informal learning behaviors: A meta-analysis. Journal of Business and Psychology, 33, 203–230. https://doi.org/10.1007/s10869-017-9492-y First citation in articleCrossrefGoogle Scholar

  • Cheung, M. W.-L. (2014). Modeling dependent effect sizes with three-level meta-analyses: A structural equation modeling approach. Psychological Methods, 19(2), 211–229. https://doi.org/10.1037/a0032968 First citation in articleCrossrefGoogle Scholar

  • Clow, D. (2013). An overview of learning analytics. Teaching in Higher Education, 18(6), 683–695. https://doi.org/10.1080/13562517.2013.827653 First citation in articleCrossrefGoogle Scholar

  • Cohen, J. (1960). A coefficient of agreement for nominal scales. Educational and Psychological Measurement, 20(1), 37–46. https://doi.org/10.1177/001316446002000104 First citation in articleCrossrefGoogle Scholar

  • Conijn, R., Snijders, C., Kleingeld, A., & Matzat, U. (2017). Predicting student performance from LMS data: A comparison of 17 blended courses using Moodle LMS. IEEE Transactions on Learning Technologies, 10(1), 17–29. https://doi.org/10.1109/TLT.2016.2616312 First citation in articleCrossrefGoogle Scholar

  • Dawson, S., Gašević, D., Siemens, G., & Joksimovic, S. (2014). Current state and future trends: A citation network analysis of the learning analytics field. Proceedings of the Fourth International Conference on Learning Analytics and Knowledge – LAK ‘14, 231–240. https://doi.org/10.1145/2567574.2567585 First citation in articleCrossrefGoogle Scholar

  • De’, R., Pandey, N., & Pal, A. (2020). Impact of digital surge during Covid-19 pandemic: A viewpoint on research and practice. International Journal of Information Management, 55, Article 102171. https://doi.org/10.1016/j.ijinfomgt.2020.102171 First citation in articleCrossrefGoogle Scholar

  • Deterding, S., Dixon, D., Khaled, R., & Nacke, L. (2011). From game design elements to gamefulness: Defining “gamification”. Proceedings of the 15th International Academic MindTrek Conference: Envisioning Future Media Environments, 9–15. https://doi.org/10.1145/2181037.2181040 First citation in articleCrossrefGoogle Scholar

  • Dunn, D. S., Saville, B. K., Baker, S. C., & Marek, P. (2013). Evidence‐based teaching: Tools and techniques that promote learning in the psychology classroom. Australian Journal of Psychology, 65(1), 5–13. https://doi.org/10.1111/ajpy.12004 First citation in articleCrossrefGoogle Scholar

  • Engeström, Y. (1987). Learning by expanding: An activity-theoretical approach to developmental research [Doctoral dissertation (monograph)]. Orienta-Konsultit. First citation in articleGoogle Scholar

  • Gašević, D., Dawson, S., Rogers, T., & Gasevic, D. (2016). Learning analytics should not promote one size fits all: The effects of instructional conditions in predicting academic success. The Internet and Higher Education, 28, 68–84. https://doi.org/10.1016/j.iheduc.2015.10.002 First citation in articleCrossrefGoogle Scholar

  • Gašević, D., Dawson, S., & Siemens, G. (2015). Let’s not forget: Learning analytics are about learning. TechTrends, 59(1), 64–71. https://doi.org/10.1007/s11528-014-0822-x First citation in articleCrossrefGoogle Scholar

  • Gnambs, T., & Appel, M. (2018). Narcissism and social networking behavior: A meta‐analysis. Journal of Personality, 86(2), 200–212. https://doi.org/10.1111/jopy.12305 First citation in articleCrossrefGoogle Scholar

  • Gurevitch, J., Koricheva, J., Nakagawa, S., & Stewart, G. (2018). Meta-analysis and the science of research synthesis. Nature, 555(7695), 175–182. https://doi.org/10.1038/nature25753 First citation in articleCrossrefGoogle Scholar

  • Hamari, J. (2017). Do badges increase user activity? A field experiment on the effects of gamification. Computers in Human Behavior, 71, 469–478. https://doi.org/10.1016/j.chb.2015.03.036 First citation in articleCrossrefGoogle Scholar

  • Harrer, M., Cuijpers, P., Furukawa, T. A., & Ebert, D. D. (2019). Doing meta-analysis in R: A hands-on guide. https://doi.org/10.5281/zenodo.2551803 First citation in articleGoogle Scholar

  • Hathaway, K. L. (2014). An application of the seven principles of good practice to online courses. Research in Higher Education Journal, 22, 1–12. https://eric.ed.gov/?id=EJ1064101 First citation in articleGoogle Scholar

  • Henderikx, M. A., Kreijns, K., & Kalz, M. (2017). Refining success and dropout in massive open online courses based on the intention – behavior gap. Distance Education, 38(3), 353–368. https://doi.org/10.1080/01587919.2017.1369006 First citation in articleCrossrefGoogle Scholar

  • Henrie, C. R., Bodily, R., Larsen, R., & Graham, C. R. (2018). Exploring the potential of LMS log data as a proxy measure of student engagement. Journal of Computing in Higher Education, 30, 344–362. https://doi.org/10.1007/s12528-017-9161-1 First citation in articleCrossrefGoogle Scholar

  • Higgins, J. P. T., Thompson, S. G., Deeks, J. J., & Altman, D. G. (2003). Measuring inconsistency in meta-analyses. British Medical Jorunal, 327(7414), 557–560. https://doi.org/10.1136/bmj.327.7414.557 First citation in articleCrossrefGoogle Scholar

  • Huang, B., & Hew, K. F. (2015). Do points, badges and leaderboard increase learning and activity: A quasi-experiment on the effects of gamification. Proceedings of the 23rd International Conference on Computers in Education, 275–280. https://www.researchgate.net/profile/Khe-Hew/publication/286001811_Do_points_badges_and_leaderboard_increase_learning_and_activity_A_quasi-experiment_on_the_effects_of_gamification/links/5665404708ae15e746333d22/Do-points-badges-and-leaderboard-increase-learning-and-activity-A-quasi-experiment-on-the-effects-of-gamification.pdf First citation in articleGoogle Scholar

  • Jo, I.-H., Kim, D., & Yoon, M. (2015). Constructing proxy variables to measure adult learners’ time management strategies in LMS. Journal of Educational Technology & Society, 18(3), 214–225. https://www.jstor.org/stable/jeductechsoci.18.3.214 First citation in articleGoogle Scholar

  • Kaufmann, E., Reips, U.-D., & Maag Merki, K. (2016). Avoiding methodological biases in meta-analysis: Use of online versus offline individual participant data (IPD) in educational psychology. Zeitschrift für Psychologie, 224(3), 157–167. https://doi.org/10.1027/2151-2604/a000251 First citation in articleLinkGoogle Scholar

  • Kebritchi, M., Lipschuetz, A., & Santiague, L. (2017). Issues and challenges for teaching successful online courses in higher education: A literature review. Journal of Educational Technology Systems, 46(1), 4–29. https://doi.org/10.1177/0047239516661713 First citation in articleCrossrefGoogle Scholar

  • Kinshuk, & Chen, N.-S. (2006). Synchronous methods and applications in e-learning. Campus-Wide Information Systems, 23(3). https://doi.org/10.1108/cwis.2006.16523caa.001 First citation in articleCrossrefGoogle Scholar

  • Knight, S. (2020). Augmenting assessment with learning analytics. In M. BearmanP. DawsonR. AjjawiJ. TaiD. BoudEds., Re-imagining university assessment in a digital world (Vol. 7, pp. 129–145). Springer International Publishing. https://doi.org/10.1007/978-3-030-41956-1_10 First citation in articleGoogle Scholar

  • Kroehne, U., & Goldhammer, F. (2018). How to conceptualize, represent, and analyze log data from technology-based assessments? A generic framework and an application to questionnaire items. Behaviormetrika, 45(2), 527–563. https://doi.org/10.1007/s41237-018-0063-y First citation in articleCrossrefGoogle Scholar

  • Kupczynski, L., Gibson, A. M., Ice, P., Richardson, J., & Challoo, L. (2011). The impact of frequency on achievement in online courses: A study from a South Texas university. Journal of Interactive Online Learning, 10(3), 141–149. http://www.ncolr.org/jiol/issues/pdf/10.3.3.pdf First citation in articleGoogle Scholar

  • Lauría, E. J. M., Baron, J. D., Devireddy, M., Sundararaju, V., & Jayaprakash, S. M. (2012). Mining academic data to improve college student retention: An open source perspective. Proceedings of the 2nd International Conference on Learning Analytics and Knowledge, 139–142. https://doi.org/10.1145/2330601.2330637 First citation in articleCrossrefGoogle Scholar

  • LeBreton, J. M., & Senter, J. L. (2008). Answers to 20 questions about interrater reliability and interrater agreement. Organizational Research Methods, 11(4), 815–852. https://doi.org/10.1177/1094428106296642 First citation in articleCrossrefGoogle Scholar

  • Lee, J., & Martin, L. (2017). Investigating students’ perceptions of motivating factors of online class discussions. The International Review of Research in Open and Distributed Learning, 18(5), 148–172. https://doi.org/10.19173/irrodl.v18i5.2883 First citation in articleCrossrefGoogle Scholar

  • Liu, Y., & Feng, H. (2011). An empirical study on the relationship between metacognitive strategies and online-learning behavior & test achievements. Journal of Language Teaching and Research, 2(1), 183–187. https://doi.org/10.4304/jltr.2.1.183-187 First citation in articleCrossrefGoogle Scholar

  • Looyestyn, J., Kernot, J., Boshoff, K., Ryan, J., Edney, S., & Maher, C. (2017). Does gamification increase engagement with online programs? A systematic review. PLoS One, 12(3), Article e0173403. https://doi.org/10.1371/journal.pone.0173403 First citation in articleCrossrefGoogle Scholar

  • Lust, G., Juarez Collazo, N. A., Elen, J., & Clarebout, G. (2012). Content Management Systems: Enriched learning opportunities for all? Computers in Human Behavior, 28(3), 795–808. https://doi.org/10.1016/j.chb.2011.12.009 First citation in articleCrossrefGoogle Scholar

  • Macfadyen, L. P., & Dawson, S. (2010). Mining LMS data to develop an “early warning system” for educators: A proof of concept. Computers & Education, 54(2), 588–599. https://doi.org/10.1016/j.compedu.2009.09.008 First citation in articleCrossrefGoogle Scholar

  • McCuaig, J., & Baldwin, J. (2012). Identifying successful learners from interaction behaviour. Proceedings of the 5thInternational Conference on Educational Data Mining, 160–163. https://eric.ed.gov/?id=ED537220 First citation in articleGoogle Scholar

  • Means, B., Toyama, Y., Murphy, R., & Baki, M. (2013). The effectiveness of online and blended learning: A meta-analysis of the empirical literature. Teachers College Record, 115, 1–47. https://agronomy.unl.edu/online/documents/Effectiveness_of_online_learning.pdf First citation in articleCrossrefGoogle Scholar

  • Mödritscher, F., Andergassen, M., & Neumann, G. (2013). Dependencies between e-learning usage patterns and learning results. Proceedings of the 13th International Conference on Knowledge Management and Knowledge Technologies – i-Know ‘13, 24, 1–8. https://doi.org/10.1145/2494188.2494206 First citation in articleGoogle Scholar

  • Mora, A., Riera, D., González, C., & Arnedo-Moreno, J. (2017). Gamification: A systematic review of design frameworks. Journal of Computing in Higher Education, 29, 516–548. https://doi.org/10.1007/s12528-017-9150-4 First citation in articleCrossrefGoogle Scholar

  • Mwalumbwe, I., & Mtebe, J. S. (2017). Using learning analytics to predict students’ performance in moodle learning management system: A case of Mbeya University of Science and Technology. The Electronic Journal of Information Systems in Developing Countries, 79(1), 1–13. https://doi.org/10.1002/j.1681-4835.2017.tb00577.x First citation in articleCrossrefGoogle Scholar

  • Nosek, B. A., Alter, G., Banks, G. C., Borsboom, D., Bowman, S. D., Breckler, S. J., Buck, S., Chambers, C. D., Chin, G., Christensen, G., Contestabile, M., Dafoe, A., Eich, E., Freese, J., Glennerster, R., Goroff, D., Green, D. P., Hesse, B., Humphreys, M., … Yarkoni, T. (2015). Promoting an open research culture. Science, 348(6242), 1422–1425. https://doi.org/10.1126/science.aab2374 First citation in articleCrossrefGoogle Scholar

  • Palvia, S., Aeron, P., Gupta, P., Mahapatra, D., Parida, R., Rosner, R., & Sindhi, S. (2018). Online education: Worldwide status, challenges, trends, and implications. Journal of Global Information Technology Management, 21(4), 233–241. https://doi.org/10.1080/1097198X.2018.1542262 First citation in articleCrossrefGoogle Scholar

  • Papamitsiou, Z., & Economides, A. A. (2014). Learning analytics and educational data mining in practice: A systematic literature review of empirical evidence. Journal of Educational Technology & Society, 17(4), 49–64. https://www.jstor.org/stable/10.2307/jeductechsoci.17.4.49 First citation in articleGoogle Scholar

  • Parry, D. A., Davidson, B. I., Sewall, C. J. R., Fisher, J. T., Mieczkowski, H., & Quintana, D. S. (2021). A systematic review and meta-analysis of discrepancies between logged and self-reported digital media use. Nature Human Behaviour, 5, 1535–1547. https://doi.org/10.1038/s41562-021-01117-5 First citation in articleCrossrefGoogle Scholar

  • R Core Team. (2021). R: A language and environment for statistical computing. R Foundation for Statistical Computing. https://www.R-project.org/ First citation in articleGoogle Scholar

  • Ransdell, S., & Gaillard-Kenney, S. (2009). Blended learning environments, active participation, and student success. The Internet Journal of Allied Health Sciences and Practice, 7(1), 1–4. https://nsuworks.nova.edu/ijahsp/vol7/iss1/9/ First citation in articleGoogle Scholar

  • Richardson, M., Abraham, C., & Bond, R. (2012). Psychological correlates of university students’ academic performance: A systematic review and meta-analysis. Psychological Bulletin, 138(2), 353–387. https://doi.org/10.1037/a0026838 First citation in articleCrossrefGoogle Scholar

  • Riley, R. D., Lambert, P. C., & Abo-Zaid, G. (2010). Meta-analysis of individual participant data: Rationale, conduct, and reporting. British Medical Journal, 340, Article c221. https://doi.org/10.1136/bmj.c221 First citation in articleCrossrefGoogle Scholar

  • Rodgers, M. A., & Pustejovsky, J. E. (2021). Evaluating meta-analytic methods to detect selective reporting in the presence of dependent effect sizes. Psychological Methods, 26(2), 141–160. https://doi.org/10.1037/met0000300 First citation in articleCrossrefGoogle Scholar

  • Saqr, M., Fors, U., & Tedre, M. (2017). How learning analytics can early predict under-achieving students in a blended medical education course. Medical Teacher, 39(7), 757–767. https://doi.org/10.1080/0142159X.2017.1309376 First citation in articleCrossrefGoogle Scholar

  • Scheffel, M., Drachsler, H., Stoyanov, S., & Specht, M. (2014). Quality indicators for learning analytics. Educational Technology & Society, 17(4), 117–132. https://www.jstor.org/stable/jeductechsoci.17.4.117 First citation in articleGoogle Scholar

  • Seifert, A., Hofer, M., & Allemand, M. (2018). Mobile data collection: Smart, but not (yet) smart enough. Frontiers in Neuroscience, 12, Article 971. https://doi.org/10.3389/fnins.2018.00971 First citation in articleCrossrefGoogle Scholar

  • Shrout, P. E., & Fleiss, J. L. (1979). Intraclass correlations: Uses in assessing rater reliability. Psychological Bulletin, 86(2), 420–428. https://doi.org/10.1037/0033-2909.86.2.420 First citation in articleCrossrefGoogle Scholar

  • Song, D. (2018). Learning analytics as an educational research approach. International Journal of Multiple Research Approaches, 10(1), 102–111. https://doi.org/10.29034/ijmra.v10n1a6 First citation in articleCrossrefGoogle Scholar

  • Song, D., & Bonk, C. J. (2016). Motivational factors in self-directed informal learning from online learning resources. Cogent Education, 3(1), Article 1205838. https://doi.org/10.1080/2331186X.2016.1205838 First citation in articleCrossrefGoogle Scholar

  • Song, D., Rice, M., & Oh, E. Y. (2019). Participation in online courses and interaction with a virtual agent. The International Review of Research in Open and Distributed Learning, 20(1), 43–62. https://doi.org/10.19173/irrodl.v20i1.3998 First citation in articleCrossrefGoogle Scholar

  • Strang, K. D. (2016). Do the critical success factors from learning analytics predict student outcomes? Journal of Educational Technology Systems, 44(3), 273–299. https://doi.org/10.1177/0047239515615850 First citation in articleCrossrefGoogle Scholar

  • Tempelaar, D. T., Rienties, B., & Giesbers, B. (2015). In search for the most informative data for feedback generation: Learning analytics in a data-rich context. Computers in Human Behavior, 47, 157–167. https://doi.org/10.1016/j.chb.2014.05.038 First citation in articleCrossrefGoogle Scholar

  • Tempelaar, D. T., Rienties, B., & Nguyen, Q. (2019). Learning engagement, learning outcomes and learning gains: Lessons from LA. Proceedings of the 16th International Conference on Cognition and Exploratory Learning in Digital Age (CELDA 2019), 257–264. https://doi.org/10.33965/celda2019_201911L032 First citation in articleCrossrefGoogle Scholar

  • Uijl, S., Filius, R., & Ten Cate, O. (2017). Student interaction in small private online courses. Medical Science Educator, 27, 237–242. https://doi.org/10.1007/s40670-017-0380-x First citation in articleCrossrefGoogle Scholar

  • Valentine, J. C., & Cooper, H. (2008). A systematic and transparent approach for assessing the methodological quality of intervention effectiveness research: The Study Design and Implementation Assessment Device (Study DIAD). Psychological Methods, 13(2), 130–149. https://doi.org/10.1037/1082-989X.13.2.130 First citation in articleCrossrefGoogle Scholar

  • Viechtbauer, W. (2010). Conducting meta-analyses in R with the metafor package. Journal of Statistical Software, 36(3), 1–48. https://doi.org/10.18637/jss.v036.i03 First citation in articleCrossrefGoogle Scholar

  • Viechtbauer, W., & Cheung, M. W.-L. (2010). Outlier and influence diagnostics for meta-analysis. Research Synthesis Methods, 1(2), 112–125. https://doi.org/10.1002/jrsm.11 First citation in articleCrossrefGoogle Scholar

  • Winne, P. H. (2004). Students’ calibration of knowledge and learning processes: Implications for designing powerful software learning environments. International Journal of Educational Research, 41(6), 466–488. https://doi.org/10.1016/j.ijer.2005.08.012 First citation in articleCrossrefGoogle Scholar

  • Winne, P. H. (2010). Improving measurements of self-regulated learning. Educational Psychologist, 45(4), 267–276. https://doi.org/10.1080/00461520.2010.517150 First citation in articleCrossrefGoogle Scholar

  • Winne, P. H. (2020). Construct and consequential validity for learning analytics based on trace data. Computers in Human Behavior, 112, 106457. https://doi.org/10.1016/j.chb.2020.106457 First citation in articleCrossrefGoogle Scholar

  • You, J. W. (2016). Identifying significant indicators using LMS data to predict course achievement in online learning. Internet and Higher Education, 29, 23–30. https://doi.org/10.1016/j.iheduc.2015.11.003 First citation in articleCrossrefGoogle Scholar

  • Zheng, L., Zhang, X., & Gyasi, J. F. (2019). A literature review of features and trends of technology-supported collaborative learning in informal learning settings from 2007 to 2018. Journal of Computers in Education, 6(4), 529–561. https://doi.org/10.1007/s40692-019-00148-2 First citation in articleCrossrefGoogle Scholar