Technology integration in postsecondary education: A summary of findings from a set of related meta-analyses

Although the overall research literature on the application of educational technologies to classroom instruction tends to favor their use over their non-use, these results vary considerably depending on what kind of technology is used, who it is used with and, more importantly, under what circumstances and for what instructional purposes it is used. Relatively recent, but well-developed and powerful methodology of systematic reviews, particularly quantitative syntheses (also known as meta-analyses) is especially suitable for addressing questions of that type by systematically summarizing research evidence in given areas of interest in social sciences.This meta-analysis summarizes data from 674 independent primary studies that compared higher degrees of technology use in the experimental condition with less technology in the control condition, in terms of their effects on student learning outcomes in postsecondary education. The result was an overall average weighted effect size of = 0.27 (k = 879, p < .01), indicating low but significant positive effect of technology integration on learning. The follow-up analyses revealed the influence of educational technology used for cognitive support and blended learning instructional settings designed interaction treatments, and technology integration in teacher training, especially when student-centered pedagogical frameworks are used. These findings are of potentially high interest and applied value for educational practitioners, including teachers and school administrators, as well as for instructional designers and developers of educational software.


The evolving role of educational technology
Like it or not, reliance on computers in various aspects of daily life is a reality and impacts significantly on nearly everything we do, both personally and professionally. Why then are educational researchers still engaging in debate over its effectiveness for teaching and learning? Some, like [11], have insisted on a rather auxiliary role for educational technology (i. e., it's functions are not unique, but can be performed in more conventional or effective ways), while others [e. g., 19] have argued that its role in education is more substantive and transformative. The disagreement is probably rooted in the history of educational technology itself. Originally, technology was utilized almost exclusively to deliver instructional content, and as a medium was no more effective than a human teacher, even an expert one. For instance, early studies on distributed closed-circuit television versus live teaching [10] found no differences between live teachers and televised teachers. Even the development of much more sophisticated computer tools and applications (e. g., computer-assisted learning, multi-media) did not improve student learning outcomes enough for the issue to be considered resolved unequivocally in favor of educational technology [e. g., 25]. It was the arrival of what [17] refers to as computer-based cognitive tools that appears to have tipped the scales from what can be achieved using multiple alternative media (i. e., argument of [11]) to what can be achieved primarily through computing (i. e., argument of [19]). Computer-based communication, simulations, serious games, blogs and wikis, social networking, search and retrieval, and the like, promise unique benefits that go well beyond the simple transfer of content from teacher to student. Add productivity software like spreadsheets, statistical packages, concept mapping programs, and a host of other student-oriented applications, and we can see that Clark's arguments [11], while still valid in certain computing domains, must be considered insufficient for examining the overall benefits (and potential deficits) of the introduction and the continued use of computing in education.

Examining the big picture -The Schmid et al. meta-analysis
While primary sample-based research is intended to provide an inferential link between a sample under scrutiny and the population that it is presumed to represent, meta-analysis makes this link more explicit by examining the effects of a multitude of primary studies that have attempted to address the same question. Inferential statistics and associated assessment of significance are unnecessary because a meta-analysis investigates the entire population of primary studies with like characteristics and conducted within a specified timeframe. In this sense, meta-analysis looks at the "big picture, " or the characteristics of the phenomenon in terms of the body of primary studies that have examined it.
The [24] meta-analysis, which forms the basis for the various adaptations that are presented here, attempts to examine the big picture in terms of learning via technology in postsecondary classrooms. Recognizing the limitations of the technology vs. no technology hypothesis (i. e., comparisons between technologyenhanced and technology-free classrooms), [24] divided the entire collection of included studies into those with no technology in the control condition and those with technology in the treatment condition and some technology in the control condition. The timeframe was 1990 through 2010 -twenty years of research on the use of technology in postsecondary classrooms.
The [24] meta-analysis followed from research that also examined the big picture, but where the unit of analysis was the meta-analysis, not the primary study [27]. In this second-order meta-analysis, 25 previously published meta-analyses that cut across all levels of formal education, subject matters and technology types, from the 1970s to the present, were selected from a pool of about 75 meta-analyses and their results synthesized. The analysis revealed a weighted average effect size of g = 0.35 (p < .01), encompassing 1,055 primary studies and 109,700 participants.

287
РОССИЙСКИЙ ПСИХОЛОГИЧЕСКИЙ ЖУРНАЛ • 2016 ТОМ 13 № 4 The [27] addressed the big question based on meta-analyses that looked at the technology vs. no technology question -the kind that [11] decried as flawed and full of confounds. The authors concluded that, generally speaking, technology does enhance learning, even if only to a relatively small extent.
One of the problems with a second-order meta-analysis, however, as pointed out by [13], is that it represents a limited means for addressing the host of peripheral questions that can only be settled in a primary meta-analysis, where coding decisions can be made by the meta-analyst and the synthesis is conducted at the more granular level of the individual effect size. As a result, [27] did not examine any substantive moderator variables and thus provided little texture or nuance to the overall results.
It is these issues, plus the fact that the educational research landscape is rapidly changing to include many new educational technology applications and relatively recent "technology vs. technology" types of experimental settings (i. e., a more reasonable question in modern education), that motivated the effort by [24] to perform a primary meta-analysis of technology use in postsecondary classrooms.
The [24] meta-analysis reports the overall weighted average effects of technology use on achievement and attitude outcomes and explores a fairly large set of moderator variables in an attempt to explain how technology treatments lead to positive or negative effects when educational technology is broadly understood in terms of the earlier definition by [23] as a "…variety of modalities, tools, and strategies for learning, [whose] effectiveness, therefore, depends on how well [they] help teachers and students achieve the desired instructional goals" (p. 19).
Out of an initial pool of 11,957 study abstracts, 1,105 were chosen for analysis, yielding 879 achievement effect sizes after pre-experimental designs and studies with obvious methodological confounds were removed. The random effects weighted average effect size for achievement was g = 0.27, k = 879, p < .05. The collection of achievement outcomes was divided into two sub-collections, according to the amount of technology integration in the control condition. These were "no technology" in the control condition ( g = 0.25, k = 479, p < .01) and "some technology" (though necessarily a lower degree of its use than in the treatment condition) in the control condition ( g = 0.31, k = 400, p < .01). Random effects multiple meta-regression analysis was run on each sub-collection revealing three significant predictors (subject matter, degree of difference in technology use between the treatment and the control, and pedagogical uses of technology). The set of predictors for each sub-collection was both significant and homogeneous. Differences were found among the levels of all three moderators, but particularly among varieties of cognitive support applications. These findings are presented in greater detail in the Results section.

288
RUSSIAN PSYCHOLOGICAL JOURNAL • 2016 VOL. 13 # 4 In summary, the [24] meta-analysis was intended to: 1) overcome some of the limitations of the previous reviews, including [27]; 2) provide the most comprehensive first-order meta-analysis of technology use in postsecondary classroom education; 3) encompass studies with "no technology in the control condition" and studies with "various degrees of technology in each condition"; and 4) look at the pedagogical use of a broad range of educational technologies and other important instructional moderator variables.
Examining the details: Three follow-up meta-analyses A set of subsequent (follow-up) analyses addressed several additional questions we judged to be of the utmost importance. More specifically, we investigated the effects of various purposes of technology use (with the focus on cognitive support for learning), technology use in blended learning, instructional settings (interaction treatments) designed specifically to enable student collaborative work, and various pedagogical approaches to teacher education. Following is a brief rationale for each of these follow-ups.

Major purpose of technology use
As educational technology advanced beyond media whose primary role was to deliver content to students (e. g., instructional television, multi-media applications, computer-assisted instruction), the following question arose in the theoretical and practical literature: "How can computers be used to support student cognition, without directly instructing them"? The [17] addresses the practical issue, but [12] provides a possible answer to the theoretical question. He argues that the most compelling role of computing in learning is its ability to afford "cognitive efficiencies" to students. In most, if not all, learning situations there is shared cognition among the learner, the task itself, and the tools that the learner uses in the process. Computer-based cognitive tools, he argues, could be designed and implemented to fulfill one of these roles. His rationale was that the more learners can distribute cognition "outside of their heads, " the more cognition can be devoted to the process of learning new material. Therefore, one purpose of the [24] review was to explore whether and to what extent cognitive tools promote student achievement in learning environments involving technology, compared to other roles that computers might assume (e. g., presentation, search and retrieval, communication).

The effects of blended learning
There is a growing literature of studies investigating blended learning, an instructional approach that involves a combination of elements of face-to-face and online instruction. It is sometimes argued that blended learning is the "best of both worlds" because it is a marriage of the best elements of the two practices, although it is still arguable what these "best elements" are and how they fit together. To date, there have only been three meta-analyses devoted to blended learning [3 -partly reported here, 21, and 26].
In the [21] meta-analysis, blended learning conditions were found to significantly outperform fully face-to-face classroom instruction ( g = 0.35, k = 23, p = .001). Several moderator variables were significant: 1) blended learning outperformed self-study; 2) collaborative learning and teacher-directed expository instruction significantly outperformed self-study; 3) in computer-mediated communications with instructor and among students, the asynchronous mode only was more effective than in combination with the synchronous mode; and 4) undergraduate students benefited more from blended learning than graduate students.
In the [26] study, the overall effect size favored blended learning ( g = 0.34, k = 24, p < .05) for achievement based on objective effectiveness outcomes and ( g = 0.34, k = 11, p < .05) for achievement based on subjective effectiveness outcomes. These results closely mirror the [21] results. The presence or absence of quizzes appeared to differentiate the overall result in terms of objective effectiveness.
The [3] study, an offshoot of [24], addressed the following research questions: What is the impact of blended learning (i. e., courses that combine face-to-face and online learning) on the achievement of higher education students? How do various pedagogical factors (e. g., the amount of time spent online and the purpose of technology use) and course demographics (e. g., subject matter) moderate the overall average effect size? These findings appear in the Results section.

Designed interaction treatments
Institutions of higher education provide students with the opportunity to interact with each other both inside and outside the classroom. Three major forms of interaction are important for effective learning, extrapolated from the distance education literature [e. g., 1,22]. Student-student interaction refers to interaction among individual students or among students working in small groups. Student-teacher interaction traditionally focused on, but is not limited to, classroom-based dialogue between students and the instructor. Finally, student-content interaction refers to students interacting with the subject matter under study to construct meaning, relate it to personal knowledge, and apply it to problem solving. Meta-analytical findings support the overall positive influence of the three types of interaction on learning outcomes with specific emphasis on student-student interactions [4]. Naturally, all three of these forms can occur in higher education classrooms without technology, but it was the intention of this study to investigate student-student interaction, supported by technology, in conditions that were specifically designed to support collaborative learning. Thus, the research question became: Are designed interaction treatments (i. e., intentionally implemented collaborative instructional conditions that are meant to increase student learning) more effective than contextual interaction treatments (i. e., learning setups that contain conditions for student-student interaction to occur, but are not intentionally designed to create collaborative learning environments in technology-enhanced classroom instruction in higher education)? This question is important because it emphasizes the use of instructional design to facilitate the potentially positive effects of technology use in higher education classrooms [7].

Technology use in teacher education
This section focuses on the subset of studies that specifically addressed technology use in educational and teacher training programs. The main objective is to further explore the impact of technology use in this specific context in an attempt to determine what aspects of teaching practices set education apart from STEM and non-STEM disciplines. Moreover, it aims to investigate the nature of the most effective pedagogical frameworks supporting successful technology integration. The focus is on how technology is used by educational professionals to achieve educational goals.
While it is impossible to review all pedagogical frameworks available to instructors, it seems appropriate to focus on student-centered instructional strategies. It is argued that the learner-centered approach supports learning by emphasizing the student's role in the instructional environment, thus shifting the focus from knowledge transmission to the actual learning process [e. g., 20]. Regardless of technology integration, general teaching strategies that are aligned with the student-centered philosophy include cooperative and collaborative learning [9,16], problem-based learning [18], and the provision of elaborate feedback [14].

Research questions
To summarize, the meta-analysis under consideration and its follow-ups were designed to answer questions about the impact of instructional technology on postsecondary student achievement outcomes. Specifically, the research questions addressed were as follows: What is the weighted average effect size and variability for studies that investigate the impact of the instructional uses of technology on postsecondary student achievement outcomes?
Is there a difference in average effect sizes for achievement outcomes associated with major purposes of technology use?
What is the weighted average effect size and variability for studies that investigate the impact of the instructional uses of technology in so-called blended educational settings?
Is there a difference in average effect sizes for achievement outcomes associated with designed vs. contextual interaction treatments?
What is the effectiveness of instructional technology in teacher training dependent on a particular pedagogical approach?

Method
To facilitate navigation through the review, here is a brief summary of major terms and definitions used in the meta-analysis, followed by a set of inclusion/ exclusion criteria and the key aspects of the review methodology (for the complete description of the methodology, please see [24]).

Terms and definitions
Educational technology is understood here according to [23] as quoted previously. The degree of technology use was the primary determinant for assigning groups to either the experimental or the control condition. This distinction is important because it specifies the +/-valence of the effect size. Two types of studies were found, those that contained no technology in the control condition and those that contained some technology in the control condition. In the former class of studies, the assignment of the experimental and control group designation was clear. In the latter case, the differential use of technology in the two conditions was rated. The condition containing the "highest degree of technology use" was designated the treatment condition and the alternative condition was the control. There were three possible interpretations of the degree of technology use. The experimental group was considered: a) to contain the longest or most frequent exposure to technology tools; b) to contain more advanced technology (i. e., enabling more functions); and/or c) to employ a larger number of technology tools.
There were the following major purposes of technology use identified and analyzed in the reviewed studies: 1) to promote communication and/or to facilitate the exchange of information. This category includes technology that enables a higher level of interaction between individuals (i. e., two-way communications among learners and between learners and the teacher); 2) to provide cognitive support for learners. This category encompasses various technologies that enable, facilitate, and support learning by providing cognitive tools (e. g., concept maps, simulations, wikis, different forms of elaborate feedback, spreadsheets); 3) to facilitate information search and retrieval. This type of technology is intended to enable and/or facilitate access to additional information (e. g., web-links, search engines, electronic databases); 4) to enable or enhance content presentation (e. g., PowerPoint presentations, graphical visualizations, computer tutorials with limited interactive features). When more than one purpose was identified, codes indicating multiple purposes were created (e. g., cognitive support plus presentational support). Achievement outcomes included various objective measures of academic performance (e. g., exam/test scores), but not self-evaluation. A wide spectrum of moderator variables were coded: methodological (e. g., research design), instructional (e. g., purpose of technology use, pedagogical approach), demographic (e. g., subject matter), and publication (e. g., date, source).

Inclusion/exclusion criteria and review procedure
Studies under review that were to be included had to have the following characteristics: 1) be published no earlier than 1990; 2) be publicly available (or archived); 3) address the impact of computer technology on student achievement and/or attitudes; 4) contain at least one between-group comparison, where one group fits the definition of the experimental condition and the other group the definition of the control condition, using the criterion of the degree of technology use (higher vs. lower); 5) be conducted in a formal postsecondary educational setting; 6) represent classroom or blended instructional environments, but not distance education; and 7) contain sufficient statistical information for effect size extraction.
Failure to meet any of these criteria led to the exclusion of the study with the reason(s) for rejection documented for further reporting. Two researchers working independently rated the studies, first at the abstract level, then at the full text level, on a scale, from 1 (definite exclusion) to 5 (definite inclusion). All disagreements were discussed until they were resolved inviting a third opinion when necessary, and initial agreement rates calculated as Cohen's Kappa (κ) and as Pearson's r (where appropriate). Similarly, two researchers participated in all other data extraction procedures (i. e., effect size extraction and study feature coding).

Literature search strategies and data sources
Extensive literature searches were designed to identify and retrieve primary empirical studies relevant to the major research questions. Key terms used in search strategies, with some variations to account for specific retrieval sources, included: "technolog*, " "comput*, " "web-based instruction, " "online, " "Internet, " "blended learning, " "hybrid course*, " "simulation, " "electronic, " "multimedia" OR "PDAs" etc.) AND ("college*," "university," "higher education," "postsecondary," "continuing education," OR "adult learn*") AND ("learn*," "achievement*," "attitude*," "satisfaction," "perception*," OR "motivation," etc.), AND excluding "distance education" or "distance learning" in the subject field. In addition, Google Internet searches were performed to help identify gray literature, including a search for conference proceedings. Review articles and previous meta-analyses were used for branching, and the tables of contents of major journals in the field of educational technology (e. g., Educational Technology Research & Development) were manually searched.

Effect size calculation and synthesis
A d-type standardized mean difference effect size was used as the common metric (i. e., Cohen's d), and then was transformed into Hedges' g metric [15] to provide necessary correction for small sample sizes. The random effects model [6] was the main analytical approach for this meta-analysis. A mixed effects model was used to test the difference in levels of moderator variables. In a mixed analysis, average effect sizes for categories of the moderator are calculated using a random effects model. The variance component Q-Between is calculated across categories using a fixed effect model [6]. All analyses, including sensitivity and publication bias analysis, were performed in Comprehensive Meta-Analysis™ 2.2.048 [5].

Results
The findings of the meta-analysis of effects of classroom technology integration in higher education on student achievement outcomes are presented, first, overall, and then by individual research sub-question as outlined earlier. More detailed information regarding each of these follow-up meta-analyses can be found in respective publications.

Overall findings
The overall random-model results of the [24] study are shown in Table 1. The total of 879 effect sizes produced a weighted average effect size of 0.27 that was significantly greater than zero. The collection is significantly heterogeneous, based on findings from the fixed model where heterogeneity is tested in terms of the magnitude of Q-Total (i. e., total between-study variability). An effect size of 0.27 is considered to be small and represents a difference of 0.27sd between the mean of the treatment condition and the control condition, amounting to about an 11 % difference. These results suggest that technology-supported instruction is advantageous compared to either non-use or limited use, but that this advantage is relatively modest.

Major function of technology use
The results become differentiated when effect sizes are divided by pedagogical application. These results from [24] indicate that technology that is used to support student cognition outperforms all other categories, but especially presentational support (Table 2). This effect is interpreted as a difference primarily between "technology used by students" (for content understanding) and "technology used by teachers" (for content delivery). Other functions, such as support for communication and a mixture of cognitive and presentational support, fall in between these two.

Blended learning
The effects of blended learning (i. e., partly in class and partly online) were derived from the [24] database and analyzed and reported in [3]. As evident from Table 3, the random-effects weighted average effect size (= 0.334, k = 117) is larger than the overall average effect of technology use in the original metaanalysis and is in line with the findings from the other meta-analyses of blended learning [21,26]. Apparently, there is an advantage that accrues from balancing face-to-face instruction with online learning outside of class. The mechanisms of this effect have not been determined, so one of the challenges of educational technology research of the future will be to tease out the effects of variables such as amount of time devoted to each pattern of instruction, the most effective learning strategies, and the teacher's role in the online portion of blended learning.

Interaction treatments
Interaction treatments were defined by [4] as instructional setups in distance education that are intended to facilitate and promote interaction among students, between students and teachers, and between students and the content. This definition was used to code studies in the [24] study in classroom setting. Table 4 shows the results of this basic analysis. Conditions where interactions were greater in the treatment group, compared to the control condition, produced results that were significantly higher than when the control group was higher in the potential for interaction. As expected, when the two conditions were roughly the same (i. e., = 0.29, k = 703), the outcome was not significantly different from the former condition (i. e., = 0.34 vs. 0.29). When the interaction treatments were divided further by the presence or absence of design intention, the results strongly supported designed interaction treatments over contextual interaction treatments (Table 5). It appears that merely providing the means for student-student interaction is not enough as a pedagogical strategy. Some form of instructional design is needed (e. g., collaborative learning, reciprocal teaching). Table 5.

Technology use in education courses
In this final section, the effects of educational technology across subject areas were investigated. Because of our special interest in teacher education (i. e., teaching future teachers), these studies were broken out from the other Non-STEM content areas and their average effect size compared with that of STEM (i. e., Science, Technology, Engineering, Mathematics) and the remaining Non-STEM subject areas. The results, shown in Table 6, revealed nearly equal average effects for STEM and Non-STEM minus education, but a noticeable difference between these areas and technology used in teacher education. To investigate further what might account for such a high average effect for teacher education, studies were further classified according to the underlying pedagogical frameworks for using technology in these studies (Table 7). Only studies from the Education sub-group, where there was a clear indication of a particular pedagogical approach to instruction and associated use of technology, were included. This reduced the number of effect sizes from k = 66 to k = 39. Although this number after the split somewhat limited the power of subsequent analyses, using technology to provide feedback to students resulted in an unusually high average effect size of = 0.75 (k = 11). Two other instructional approaches to technology use were also prominent: multimedia theory and problem-based learning were each around = 0.50. Notably, the use of technology to support collaborative learning (i. e., group projects) was very low (less than = 0.10) and not significantly greater than zero. These finings may arise from the peculiarities of teaching future teachers, as compared with content, as in the other STEM and Non-STEM subject areas, but it does suggest the need for further exploration of the range of technology uses in education, as well as in allied areas such as nursing education.

Discussion
Overall results. Overall, these results demonstrate a consistent message concerning the value of technology use in higher education. Whether or not these results could have been achieved through other means, as [11] has claimed, is rather a moot point in our view. For better or for worse, technology is with us for the long haul and we as a profession and as professional researchers are obligated to analyze and investigate it, and to make certain that the use of it for pedagogical purposes achieves maximum learning benefits.
Cognitive versus presentational support tools. Based on the analyses of the purposeful use of technology, we see where Clark's original assertions about the passive nature of technology may have arisen. Technologies prior to the 1980s and well into the 1990s were indeed passive because their primary purpose was to convey content to students, either through the actions of teachers' use of presentational software or as a consequence of computer-aided-instruction (CAI). Feedback, of course, was present in these stand-alone technologies, but it often amounted to simply providing the "right answer" without elaboration [2]. Not until the advent of personal computers, and especially software and online tools that "work with" the student in the learning process, have we seen changes in technology use that educators could not have envisaged prior to 1980, when much of Clark's work was done.
The overall message emerging from these data is that learning is best supported when the student learns through meaningful activities via technological tools that provide cognitive support during the process. However, we are a long way from understanding more specifically how to design effective cognitive support tools and when precisely to integrate them into instruction. We encourage vigorous research programs to help bridge our knowledge gaps in these areas.
Blended learning in higher education. For the longest time, there were two primary venues for learning in higher education -classroom instruction (with or without technology support) and distance education, now often referred to as online learning. While there is considerable evidence that technology that is used to support classroom instruction is beneficial, it is a requirement when students and teachers are in separate locations, often working asynchronously (in the correspondence education era, the post office was a technology, of sorts). We are now able to marry these two environments, providing students with some time in classrooms and some time online out of classrooms. The results of studies of these so called blended or hybrid learning experiences are encouraging [3,21,26], but we are still unable to predict with confidence which variables are most influential (e. g., instructional strategy applied in each context) and which are trivial (e. g., time spent in each pattern), or how to design effective blended learning given the myriad of circumstances that can arise under various conditions. We must encourage research work in this domain, as it may turn out to be of immense value to university students, partly because it encourages a pattern of face-to-face and online work that they will encounter throughout their future careers.
Collaborative interaction treatments. In this meta-analysis we have identified a link between the intentioned and designed use of technology in higher education classrooms and the provision of technology without such explicit intention [7]. The results provide educators with specific guidance about what works and doesn't work in the domain of education, especially as it relates to technology-based student-student interaction. This part of the meta-analysis reaffirms the effectiveness of computer supported collaborative learning (CSCL) from the perspective of technology use and the design of instruction that aims to support and amplify interaction. Once again, pedagogy and specific instructional design take precedence over the contextual communicative benefits of modern educational technology.
Technology use in teacher education. The current analysis has moved one step beyond responding to the general question of whether technology works or not. Findings have confirmed previous results and provided meaningful insights with regard to specific pedagogical approaches that are successful in improving student performance. The general analyses were in line with the findings of the overall meta-analysis [28], indicating the importance of cognitive support tools for successful learning. In addition, the results further suggested that moderate intensity and complexity of technology use works better than oversaturation.
The current meta-analysis provides some input regarding pedagogical strategies that work better for educating pre-service teachers. Particularly speaking, provision of adequate and specific feedback to students in technologysupported environments greatly increases the impact of technology use on student performance. The resulting average effect size of 0.75 translates into a 27-percentile gain for average students in the experimental group compared to those in the control group. The gains are significant and the implications are clear. Instructors need to incorporate feedback in their technology-enhanced instruction. Another pedagogical approach for successful use of technology in educational contexts is problem-based learning (PBL), where the average effect size of 0.56 indicates a 21-percentile gain in student performance.
Finally, one of the most interesting pieces of evidence offered by this study pertains to the importance of training in successful technology integration in post-secondary education. This is in striking contrast to the belief that the newer generation of students is so technologically savvy that they do not need training.

Concluding remarks
In conclusion, we would like to specifically highlight for the readers two particular outcomes of this entire series of meta-analyses that, in our view, are of the utmost importance for research and practice: 1. Technology alone, no matter how advanced, sophisticated, and fashionable, hardly works beyond its "novelty effect" in the absence of the other operative consideration -"educational. " Well-thought through instructional design and effective pedagogical strategies (e. g., interaction treatments designed for collaborative work, elaborate feedback, in-time technology training for teachers and students) provide the substantial value-added that transforms technology (i. e., whichever one we care to adopt) into technological tools that are advantageous for teaching and learning.
2. Blended learning, which supposedly combines the best qualities of face-toface and online instruction, appears to be a viable teaching/learning option for applying educational technology to achieve its maximum benefits for learning.