Volume 38, Issue 3 p. 797-810
REGISTERED REPORT
Full Access

Determinants of student performance with mobile-based assessment systems for English as a foreign language courses

Jorge Bacca-Acosta

Corresponding Author

Jorge Bacca-Acosta

Fundación Universitaria Konrad Lorenz, Faculty of Mathematics and Engineering, Bogotá, Colombia

Correspondence

Jorge Bacca-Acosta, Fundación Universitaria Konrad Lorenz, Faculty of Mathematics and Engineering, Carrera 9Bis 62-43, Bogotá, Colombia.

Email: [email protected]

Search for more papers by this author
Ramon Fabregat

Ramon Fabregat

University of Girona, Institute of Informatics and Applications, Girona, Spain

Search for more papers by this author
Silvia Baldiris

Silvia Baldiris

Fundación Universitaria Tecnológico Comfenalco, Direction of Research, Innovation and Social Projection, Cartagena (Colombia), Universidad Internacional de la Rioja, Vice Rectory of Research, Logroño, Spain

Search for more papers by this author
Kinshuk

Kinshuk

University of North Texas, College of Information, Denton, Texas, USA

Search for more papers by this author
Juan Guevara

Juan Guevara

Universidad Distrital Francisco José de Caldas, Facultad Tecnológica, Bogotá, Colombia

Search for more papers by this author
First published: 23 January 2022
Citations: 1

Funding information: Departament d'Universitats, Recerca i Societat de la Informació, Grant/Award Number: 2017 SGR-1551; Fundación Universitaria Konrad Lorenz, Grant/Award Number: 5INV8181; Secretaría de Estado de Investigación, Desarrollo e Innovación, Grant/Award Number: TIN2014-53082-R; Universitat de Girona, Grant/Award Number: MPCUdG2016

Abstract

Background

Mobile-based assessment has been an active area of research in the field of mobile learning. Prior research has demonstrated that mobile-based assessment systems positively affect student performance. However, it is still unclear why and how these systems positively affect student performance.

Objectives

This study aims to identify the determinants of student performance during students' use of a mobile-based assessment application in a formative assessment activity as part of English as a Foreign Language courses in higher education.

Methods

A structural model based on hypotheses will be validated using partial least squares-structural equation modelling with data from the interaction of around 100 students of English as a Foreign Language (EFL) courses from the A1 and A2 levels of English that used a mobile-based assessment system for a period of 4 weeks.

Results and Conclusions

This registered report describes the related work, hypotheses development, methodology, and proposed analysis to validate the structural model based on hypotheses. No results or conclusions have been obtained yet.

Lay Description

What is currently known about this topic

  • Mobile-based assessment increase students' learning outcomes and motivation.
  • Mobile-based assessment provides some advantages with respect to other modes of assessment.
  • The determinants of mobile-based acceptance have been identified.

What this paper adds

  • Determinants of student performance during students' use of mobile-based assessment systems.
  • An automatic data collection method included in the mobile-based assessment application.
  • A more longitudinal intervention to identify a real effect of mobile-based assessment.

Implications for practice and/or policy

  • Designers of mobile-based assessment systems can be better informed to increase student performance.
  • The learning process can be personalized based on the determinants identified.
  • Suggestions for mechanisms to automatically detect changes in student performance.

1 INTRODUCTION

Assessment is a fundamental part of the learning process (Clements & Cord, 2013; Romero et al., 2009; Syarifudin & Suharjito, 2020; Yan, 2019) that can also be supported by mobile devices (Harchay et al., 2014). This field of research is often known as mobile-based assessment. Mobile-based assessment is a mode of assessment supported by mobile devices and other wireless technologies (Nikou & Economides, 2017b), and it has been an active area of research in the last decade (Harchay et al., 2014). Some mobile-based assessment systems have been successfully implemented in many different subject areas, such as environmental and engineering education, inquiry science learning, and health care, among others (Nikou & Economides, 2013). According to Nikou and Economides (2017a), the advantages of mobile-based assessment include easier administration, time and location independence, ubiquity, personalization, social interactivity, and context awareness.

There is a large body of literature that reports on the positive effects of mobile learning on student performance (Adams Becker et al., 2017). However, there is a latent need for further research on the effect of mobile learning in the classroom (Mendes et al., 2018). Research on mobile-based assessment systems has demonstrated that these systems support the assessment of authentic/situated learning activities (Santos et al., 2015), provide an instant formative assessment (Sung et al., 2016), increase students' learning outcomes when compared to other modes of assessment (paper-based and web-based) (Karadeniz, 2009; Senthil Kumaran, 2015; Syarifudin & Suharjito, 2020) and increase learning motivation, satisfaction and perceived learning performance (Nikou & Economides, 2016; Nikou & Economides, 2021; Syarifudin & Suharjito, 2020). However, according to Nikou and Economides (2018a), one of the main research gaps in the literature is that it is still unclear how mobile-based assessment systems have a positive impact on student performance. In other words, it is unclear which are the determinants of student performance during students' use of mobile-based assessment systems in the context of formative assessment activities. This is a research gap in the field of mobile learning that deserves the attention of researchers, and this study seeks to address part of this gap in the literature.

As for the mode of assessment, in this paper, we focused on formative assessment activities. According to Nikou and Economides (2013), in the field of assessment, a common distinction in the literature is between formative assessment and summative assessment. Formative assessment means that students' performance and work can be used to improve the learning process and avoid randomness and trial-and-error learning (Sadler, 1989). Moreover, formative assessment helps teachers obtain information about students' performance during the course, and this information can be used to adapt instruction (Chen & Chen, 2009). In contrast, summative assessment refers to the type of assessment in which there is a report of students' progress at the end of a learning unit and it is also called the assessment of learning (Nikou & Economides, 2013). Summative assessment judge competencies at the end of instruction (Chen & Chen, 2009).

In this context, the research question that drives this study is: Which are the determinants of student performance in formative assessment activities with mobile-based assessment systems in higher education? English as a Foreign Language (EFL) is taken as the subject domain in this study.

In short, the main contribution of this study is the identification of some predictors of student performance in formative assessment activities with a mobile-based assessment application in higher education EFL courses.

The rest of this paper is organized as follows: Section 2 describes the related work; Section 3 describes the hypotheses development. Section 4 describes the methodology and research design that will be followed and describes how to analyse the results using the partial least squares-structural equation modelling (PLS-SEM) technique. Finally, the references are provided.

2 RELATED WORK AND THEORETICAL BACKGROUND

Mobile-based assessment is a mode of assessment supported by mobile devices and other wireless technologies (Nikou & Economides, 2017b; Nikou & Economides, 2019a) and is an alternative to paper-based or computer-based assessment (Nikou & Economides, 2017a). Mobile-based assessment has been an active area of research in the last decade (Harchay et al., 2014), but it is still an emerging area of mobile learning (Nikou & Economides, 2018b; Alrfooh & Lakulu, 2020). Some advantages of mobile-based assessment over other modes of assessment such as computer-based assessment or paper-based assessment are: easier administration, automatic feedback, instant results, context-awareness, ubiquity, situatedness, personalization, time and location independence, and on-demand assessment (Kumaran, 2015; Nikou & Economides, 2017a; Nikou & Economides, 2018a). In other words, mobile-based assessment takes advantage of the mobile device's sensors to provide an increased assessment experience. Recent research has shown that, in terms of user experience, mobile-based assessment is more attractive, enjoyable, pleasing, and innovative when compared to computer-based assessment (Nikou & Economides, 2019b).

User experience, time and location independence, and context-awareness are advantages of mobile-based assessment systems over other forms of assessment such as computer-based assessment and paper-based assessment (Nikou & Economides, 2017a; Nikou & Economides, 2018a; Syarifudin & Suharjito, 2020; Yarahmadzehi & Goodarzi, 2020). In this study, these advantages (i.e., user experience, time and location independence, and context-awareness) make the mobile learning experience different from others in the literature and different from a computer-based assessment. The primary rationale behind this is that students who will participate in this study often spend more than 1 h and a half in public transport when they have to go from home to the university, and they often report using their mobile devices for learning during this time. This form of mobile learning has been reported in previous research (Schuck et al., 2017) and refers to learning ‘on the move’ (Sharples, 2013). In addition, some of the students are working full time and taking night classes. In this learning context, for these students, the use of mobile devices become an opportunity for test preparation in language learning (or other subjects) because students can use their mobile device for learning while they go from home to the university or when they return to home. Given these conditions, the mobile-based assessment can be context-aware and identify the learning context (i.e., travel on public transport, waiting at the bus stop, being at home or being at the university) to provide different practice tests. In this context, this study attempts to identify the determinants of student performance during students' use of mobile-based assessment systems under these conditions of mobility and availability of time for test preparation. Some unanswered questions are: are mobile-based assessment systems effective for test preparation when students have limited time or have to move on public transport for long periods? What are the determinants of student performance in these learning contexts?

According to Senthil Kumaran (2015), mobile-based assessment applications might be better at increasing students' learning performance compared to traditional paper-based assessment and e-Learning assessment. Mobile-based assessment has been more effective for increasing student motivation and achievement when compared to paper-based assessment (Nikou & Economides, 2016). Karadeniz (2009) found that students do not like paper-based assessment when they compare it to web-based or mobile-based assessment because feedback in paper-based assessment is delayed and general, while in mobile-based and web-based assessment students receive instant and more specific feedback. Thus far, previous studies in the field suggest that mobile-based assessment can be more effective than other forms of assessment.

Most of the prior research on mobile-based assessment systems has focused on Science, Technology, Engineering, and Mathematics (STEM) subjects. However, there is a lack of research on non-STEM subjects (Nikou & Economides, 2018b) with the exception of some studies conducted in non-STEM subjects such as work-based learning in health and social care (Taylor et al., 2010), learning about local culture (Hwang & Chang, 2011), visual communications (Chao et al., 2014) with positive results in terms of the effectiveness of mobile-based assessment for increasing student performance, acceptance and intention to use mobile-based assessment (Nikou & Economides, 2017b).

However, there is a lack of research on mobile-based assessment on language learning, and the studies conducted so far do not report the determinants of students' learning outcomes when students use mobile-based formative assessment activities. Language learning differs from learning a STEM subject in that learning a second language is not a linear process and is not uniform (Benigno et al., 2017). Skills in language learning are developed in parallel while knowledge in STEM is linear. Learning a second language also depends on contextual factors such as the possibility of practicing the language in real situations, the similarity of the first language with respect to the second language, and the interaction with other learners (Benigno et al., 2017) but some of this factors are not critical in STEM subjects. In that regard, the applicability of mobile-based assessment requires focused research. Consequently, research in mobile-based assessment in other subjects such as STEM is not directly transferable to the context of language learning. Moreover, more studies in mobile learning for assessment purposes with larger research samples (de-Marcos et al., 2010) and longitudinal studies (Sung et al., 2016) are needed to determine the real effect of mobile learning on assessment activities.

Although findings of research on mobile-based assessment systems for STEM subjects can be generalized to non-STEM subjects, there are differences in these subjects that might worth research to identify the unique affordances of mobile-based assessment systems for non-STEM subjects such as language learning. Therefore, this research focuses on language learning to contribute to the body of research on mobile-based assessment for language learning.

In the context of formative assessment, some studies have explored the combination of mobile learning and formative assessment activities. For instance, Bhati and Song (2019) introduced a model that combines dynamic learning spaces and mobile collaborative experiential learning for providing personalized formative assessment and contextualized learning in developing countries. Bikanga (2018) introduced the Mobile Learning Framework for Assessment Feedback (MLFAF), a framework for designing and developing mobile learning initiatives in the context of formative assessment for providing assessment feedback to increase students' engagement and attitudes towards feedback. Chou et al. (2017) investigated the effect of using mobile devices for teaching English language under the BYOD and found that students obtained better results in learning retention when they used a formative evaluation with the mobile device. These studies shed some light on the effect of formative assessment activities with mobile devices. However, it is still unclear what are the factors that influence student performance during formative assessment activities during students' use of mobile learning applications.

Recent research on language learning has provided evidence of the positive effect of the m-learning approach on students' learning outcomes (Persson & Nouri, 2018) as well as its affordances. For instance, Syarifudin and Suharjito (2020) combined a mobile-based assessment system with augmented reality technology for learning English and found that the mobile application increased student performance and motivation. In the same vein, Ozer and Kiliç (2018) found that their m-learning environment for learning English had a significant effect on student learning performance. Sung et al. (2016) conducted a meta-analysis of 110 studies in mobile learning and found that 69.95% of the learners that used a mobile device for learning obtained better learning outcomes compared to students who did not. In a systematic literature review, Shadiev et al. (2017) found that, in general, m-learning has a positive effect on language proficiency and creates positive attitudes towards it. These results were confirmed by a systematic literature review conducted by Persson and Nouri (2018). Research on mobile language learning has reported a large number of aspects that make mobile devices unique and distinct from other technologies. Mobile devices are particularly suitable for language learning because they allow to create not only indoor learning activities but outdoor learning activities in which students are exposed to authentic contexts for learning vocabulary or practicing other skills (Persson & Nouri, 2018).

Moreover, built-in components of mobile devices such as the camera and microphone are relevant to create unique multimedia material in learning experiences that involve sharing the device under collaborative approaches (Persson & Nouri, 2018). The use of mobile devices in language learning allows teachers to use context-aware applications so that students can learn in natural environments or access the learning material outside the classroom anytime and anywhere in other contexts or locations (ubiquitous learning) such as in public transport, the campus or a park (Persson & Nouri, 2018) and that cannot be done with other technologies. Since students are used to using social networks, mobile devices are suitable for combining the use of social networks with language learning activities such as listening to audio news recordings and sharing, in a social network, a summary of what they understood (Read et al., 2021). In general, researchers concluded that learning with mobile devices seems to be significantly more effective than traditional approaches or other technologies.

Evidence from the literature also shows that mobile-based assessment positively affects learning outcomes (Bacca-Acosta & Avila-Garzon, 2021). However, it is still unclear what are the determinants of student performance during students' use of mobile-based assessment systems for formative assessment activities. This gap in the literature motivated the study presented in this paper, and therefore this study contributes to filling part of that gap. In that regard, the next step in this study is to identify the determinants of student performance from the literature. Moreover, there are two aspects that make this study different from other research in the literature: (1) this study considers inherent processes of language learning such as the development of vocabulary, grammar, and listening in the context of a mobile-based formative assessment activity, that are not common in other subjects (i.e., STEM) and remain unclear in the literature on mobile-based assessment in language learning (2) the educational intervention is not planned to be cross-sectional but more longitudinal in nature to have a more accurate view of the effect of mobile-based assessment systems on student performance as suggested by previous research in the field (Nikou & Economides, 2016).

3 HYPOTHESES DEVELOPMENT

To identify some of the factors that might be determinants of student performance in mobile-based assessment systems, a search in the Scopus and Web of Science databases was conducted with the following search string: (‘mobile-based assessment’ OR ‘mobile learning’) AND predict*. In this section, we discuss some factors that were identified as factors that might predict student performance in mobile learning settings: scaffolding, time on-task, motivation, feedback, acceptance, and self-regulated learning. This section aims to define some hypotheses on how these factors might have a positive effect on student performance during the use of mobile-based formative assessment systems. These hypotheses formed a structural model that will be validated later with data collected from the use of a mobile-based assessment application in EFL courses in higher education.

3.1 Scaffolding

There are many definitions of the term ‘scaffolding’ and there is also a lack of precision on its definition because many researchers have used this term to denote different instructional methods (Belland, 2017). Scaffolding is a process that enables a novice learner to solve a problem or carry out a task that is outside of their current skills or ‘beyond his unassisted efforts’ (Wood et al., 1976). Scaffolding has been successfully implemented in different mobile learning applications (Zydney & Warner, 2016). For instance, Hung et al. (2013) confirmed that scaffolding in their m-learning application helped students to enhance their inquiry skills in ecology. According to Santos et al. (2015), in mobile assessment, the mobile device can be transformed into a ‘more capable peer’ when scaffolding strategies are used to assist learners during the assessment activity. Thus, one of the factors that might ensure success in a learning activity is the use of scaffolding strategies. Kim and Frick (2011) pointed out that the learning activities should be in the zone of proximal development so that the activities can be done with the support and guidance of the teacher and other educational resources.

A broader perspective on the research on scaffolding was provided by Belland et al. (2015), who conducted a meta-analysis on computer-based scaffolding and found that scaffolding has a positive influence on learning with an effect size of 0.53. Scaffolding has been used as a strategy to promote the development of higher-order thinking skills and content knowledge. However, to the best of our knowledge, none of the studies has explored the effect of scaffolding in mobile-based assessment systems during formative assessment activities.

Consequently, according to the literature, it seems that scaffolding has a positive effect on students' learning outcomes. Hence, the following hypothesis is proposed in which scaffolding is the independent variable and learning outcomes is the dependent variable:

H1a.Scaffolding has a positive and significant effect on student performance during students' use of mobile-based assessment systems.

3.2 Time on-task

Time on-task measures have been used as an accurate estimation of students' learning (Kovanović et al., 2015). Learners who practice and engage with a task have better learning outcomes than those who practice less or engage less with a task (Landers & Landers, 2014; Louw et al., 2008). Landers and Landers (2014) found that time on-task strongly predicted student performance. In their study, Rawson et al. (2017) conducted a correlational study and found that there is a positive correlation between the amount of time that students spend doing the homework and their learning outcomes.

Taylor et al. (2010) concluded that students spent more time using the assessment system in a mobile application compared to the time spent on a personal computer. Moreover, students spent more time per question on a mobile device than on the personal computer. Larabee et al. (2014) found that students spend more time on reading tasks when using an iPad compared to traditional learning materials. Together these studies suggest that time on-task positively influences student performance when technology is used to present the content or mediate the learning process. Together these results provide evidence of the positive effect that time on-task has on student performance. Other researchers have concluded that time on-task influences learning performance when other variables are considered together (i.e., the interaction between other variables and time on-task; Nonis & Hudson, 2006). In that regard, some studies in the literature have reported a positive effect of scaffolding on students' time on task (Ibanez et al., 2016). This means that time on-task might be a mediating variable that mediates the relationship between scaffolding and student performance. Thus, the following hypotheses are suggested:

H1b.Scaffolding has a positive and significant effect on students' time on-task during students' use of mobile-based assessment systems.

H2.Time on-task has a positive and significant effect on student performance during students' use of mobile-based assessment systems.

3.3 Motivation

Motivation is a psychological construct that explains why people make an effort to pursue a goal and why people actively work to attain that goal (Keller, 2010). Motivation is a complex psychological construct that can be studied from a wide variety of dimensions. Consequently, we focused on some specific dimensions of student motivation that are more representative in test taking such as effort, importance, test anxiety, and pressure/tension.

3.3.1 Effort and importance

According to Schüttpelz-Brauns et al. (2018), students' learning outcomes in tests depend not only on students' abilities but also on test-taking effort. Cole et al. (2008) found that students who report high efforts on test-taking obtain better learning outcomes than those who do not. In general, there is a positive relationship between the efforts devoted to test-taking and student performance. This means, if students perceive the importance of a test, their efforts will be higher and their learning outcomes will be better (Cole et al., 2008). Consequently, a positive relationship can be established between effort and importance (as a construct of student motivation and as the independent variable) and student performance as the dependent variable. In that regard, the following hypothesis is proposed:

H3.Students' perceptions of importance of a test (as a dimension of motivation) has a positive and significant effect on student performance during students' use of mobile-based assessment systems.

H4.Students' perceptions of effort (as a dimension of motivation) has a positive and significant influence on student performance during students' use of mobile-based assessment systems.

3.4 Feedback

Feedback is defined ‘as information provided by an agent (e.g., teacher, peer, book, parent, self, experience) regarding aspects of one's performance or understanding’ (Hattie & Timperley, 2007, p. 81). Feedback is considered to be one of the most critical factors that positively influence students' learning outcomes (Abdurrahman et al., 2018; Hattie & Timperley, 2007). Feedback is a crucial aspect of formative assessment (Gedye, 2010) which in turn improves student performance (Hwang & Chang, 2011). Feedback provided in ongoing assessment significantly improves student performance and increases motivation which at the same time improves their cognitive processes (Abdurrahman et al., 2018). Moreover, feedback helps students to recognize the gap between their desired performance (competencies) and the actual achievement (Abdurrahman et al., 2018). However, research on assessment feedback and its impact on student performance is still in its infancy (Bikanga, 2018).

Feedback helps students to be aware of their current learning status and allows them to adjust their learning strategies to achieve the learning goals. Thus, students can examine their learning behaviour (Chen & Chen, 2009). Providing individual feedback before the final exams is essential to address students' needs and help students to acquire the desired competencies (Förster et al., 2018). According to Chen and Chen (2009) and Syarifudin and Suharjito (2020), feedback from formative assessment helps students to improve their performance, and therefore feedback predicts student performance. Feedback should emphasize the effort to achieve the learning objectives (Meyer et al., 2014). In general, Förster et al. (2018) and van der Kleij et al. (2012) call for more research on the effect that some feedback mechanisms might have on students' learning outcomes. Consequently, there are still some open issues in this field. Thus, the following hypothesis is proposed in which the provision of feedback is the independent variable, and student motivation is the dependent variable:

H5.The provision of feedback in mobile-assessment systems has a positive and significant effect on student performance.

3.5 Acceptance of mobile-based assessment systems

Recent research has identified some of the factors that influence the acceptance of mobile-based assessment systems (Nikou & Economides, 2017b) and defined the mobile-based assessment acceptance model (MBAAM). Nikou and Economides (2017b) and Alrfooh and Lakulu (2020) confirmed that some sub-constructs of the technology acceptance model (TAM), such as perceived ease of use and perceived usefulness, have a positive effect on the behavioural intention to use mobile-based assessment systems. Other researchers concluded that the acceptance of e-learning has a positive and significant effect on the students' learning outcomes (Pham & Tran, 2020). In the same vein, Larmuseau et al. (2018) found that perceived usefulness (one of the constructs of the TAM) has a positive effect on the actual use of an e-learning platform and has a positive effect on the students learning performance. Likewise, Tabak and Nguyen (2013) pointed out that students who perceive that a system is easy to use are more likely to believe that the system will improve their learning performance. Chan et al. (2015) confirmed these results by including the satisfaction factor and found that satisfaction with the system also influences students' learning outcomes. Ghosh (2016) concludes that perceived usefulness has a positive effect on the usage of an e-learning platform and usage has a positive impact on students' learning outcomes. Based on the evidence in the literature, we hypothesize that the acceptance of e-learning artefacts positively affects student learning performance. However, there is still a lack of research on the attitudes and acceptance towards the use of mobile-based assessment (Al-Emran & Salloum, 2017). One question that remains unanswered is how students' acceptance of mobile-based assessment systems might have a positive influence on student performance in the context of mobile-based formative assessment activities. Thus, in this study, some sub-constructs of the MBAAM model were considered to determine how the behavioural intention to use a mobile-based assessment system might affect student performance, and consequently, we suggest the following hypotheses (based on the aforementioned model):

H6a.Perceived ease of use has a positive and significant effect on the behavioural intention to use mobile-based assessment systems.

H6b.Perceived ease of use has a positive and significant effect on the perceived usefulness of mobile-based assessment systems.

H7.Perceived usefulness has a positive and significant effect on the behavioural intention to use a mobile-assessment system.

H8.The behavioural intention to use a mobile-based assessment system has a positive and significant effect on student performance during students' use of mobile-based assessment systems.

H9.Feedback has a positive and significant effect on the perceived usefulness of mobile-based assessment systems.

H10.Feedback has a positive and significant effect on the behavioural intention to use a mobile-based assessment system.

H11.The perceptions about the user interface have a positive and significant effect on the perceived ease of use in a mobile-based assessment system.

H12.The behavioural intention to use a mobile-based assessment system has a positive and significant effect on the amount of time dedicated to use the system (time on-task).

The structural model based on all the hypotheses in this research is depicted in Figure 1.

Details are in the caption following the image
Structural model of predictors of students' learning outcomes

4 METHODOLOGY AND RESEARCH DESIGN

This study aims to identify the determinants of student performance during students' use of a mobile-based assessment application in a formative assessment activity as part of EFL courses in higher education. To the best of our knowledge, this is the first study of its kind to provide insights into the factors that might explain the effect of mobile-based formative assessment activities on student performance in EFL courses in higher education. To validate the structural model of hypotheses, in this study, the data will be collected from two sources: a mobile-based assessment application, called K-English (Bacca et al., 2019), that was developed by the authors with an automatic monitoring system, and a self-reported instrument that is described in detail in Section 4.4. By collecting data from both the self-reported measures and the automatic monitoring process (non-intrusive) in the mobile app, we gathered information that is more accurate than relying only on self-reported measures to have a better overview of how students actually use the application. The research procedure and instruments were approved by the Institutional Ethics Committee of the university where the research study will take place. The informed consent will be obtained from all participants, and the procedure will be conducted in accordance with the Declaration of Helsinki (World Medical Association, 2013).

4.1 Participants

To determine the number of participants (sample size), we followed the minimum R-squared method as suggested by Hair et al. (2017). In our structural model, we have nine arrows as the maximum number of arrows pointing at the learning performance variable; we defined a 5% probability error with a statistical power of 90% for detecting R2 values of at least 0.25. The results show that our study would require a minimum of 88 participants.

Participants will be invited from EFL courses that correspond to the A1 and A2 levels of the CEFR (Common European Framework of Reference for Languages) at university level. Participation in the research will be voluntary. Students are going to use a mobile-based assessment application developed by the authors and called K-English in the context of a formative assessment activity for a period of 4 weeks to prepare students for the Cambridge KET official exam. The mobile-based assessment application was developed to enable students to practice vocabulary, grammar, and listening from a corpus of 305 questions/items included in the mobile application. At the same time, students practiced some skills for answering six different types of questions/items that usually appear in the Cambridge KET official exam. Further details of the application are provided in Section 4.2. Almost all of the students that are going to participate in the study may belong to the undergraduate programme in psychology, which is the largest higher education programme in the university where this study will be conducted.

4.1.1 Inclusion and exclusion criteria for participants

Inclusion criteria:
  1. To be enrolled in the EFL courses at the university where this study will be conducted.
  2. Participants that own a smartphone.
  3. Participants with smartphones with any version of the Android operative system.
  4. Participants that are taking the English course for the first time.
  5. Participants should have access to the internet on their mobile device by using WI-FI or mobile data.
  6. Participants with a level of English of A1 and A2 according to the CEFR levels.
Exclusion criteria:
  1. Participants with Iphone or IOS based operative systems will be excluded because the mobile-based application was developed for Android devices.
  2. Participants that are taking the English course for second or more times.
  3. Participants with a level of English higher than A2 according to the CEFR levels.

4.2 The mobile-based assessment application

For the educational intervention, a mobile-based assessment application called K-English was developed. The application was developed with the support of EFL teachers, and the primary purpose was to provide an educational tool for a formative assessment activity in which students practice some of the types of questions (items) from the Cambridge KET (Key English Test) official exam. The app K-English enables students to practice vocabulary, grammar, and listening through six different types of questions in preparation for the Cambridge KET exam. In particular, the mobile-based assessment application covers the types of questions for practicing for parts 1, 2, 3, and 6 of the reading and writing sections of the Cambridge KET as well as parts 2 and 3 of the listening section (Cambridge., 2016). For the purposes of this research, a new module was developed for the K-Enlish application. This module uses the GPS (Global Positioning System) to identify the student's location to determine if the students are at the university, at home or moving in public transport. The student's location is used to display tests that have more related questions to the specific locations. For instance, questions that contain vocabulary about household items when students are at home or vocabulary about public transport when they are moving to or from the university. In some cases, when the GPS is not available, the application shows a pop-up asking students about their current location. Regarding data privacy, students will be informed about recording their location only for research purposes while the application is being used when they are asked to participate in this research. In total, the application has 305 questions for all of the six types of questions described earlier. These questions were reproduced from the following well-known books for learning English as a foreign language: Gray and O'sullivan (2000), Capel and Ireland (2004), and Andrew Betsis (2014) after obtaining the corresponding written permission from the publishers for this research study. More information about the publishers is provided in the acknowledgements. Each question was tagged to be adapted to the student's location.

The mobile-based assessment application was designed according to the Triple-A architecture introduced by Romero et al. (2009). Thus, the system consists of three main functional layers:
  • Assembling: This layer was implemented in a web-based application in which teachers can create the questions, and teachers are able to add those questions to the questions bank. A test in the application consists of some questions of the same type (one of the six types of questions). Using the web application, teachers can set up the number of items for each test (e.g., five questions of the same type per test). Teachers can also define the feedback that will be provided to the students for each one of the questions. This application has been used to create 305 questions from the English books.
  • Administering: This layer was implemented in the mobile application (see Figures 2 and 3). The mobile application sends and receives data to and from the web application. In this layer, the questions are shown to students, and students' answers are collected by the application to display the appropriate feedback. The items are randomly selected from an items bank in the web application and are displayed by the mobile application according to the student's location obtained from the GPS or the information provided by the user. This functionality provides flexibility to update or add questions to the items bank without reinstalling the mobile application. In terms of user experience, the user interface is easy to use because the different options in the application are identified with icons. Moreover, the app includes an onboarding mechanism so that students become familiar with the app.
Details are in the caption following the image
Main menu of the mobile-based assessment application showing two options: practice and scores/statistics
Details are in the caption following the image
Screenshot of the six types of questions that students can practice in the app
The mobile application was developed with the following modules:
  • Scaffolding: Provides support to students when they are answering the questions in the application. The scaffolding mechanism in the application was designed to be in line with the dynamic assessment features (Belland, 2017). In particular, the scaffolding mechanism was designed to draw students' attention to the most essential concepts in each question and to contribute to the formative assessment process. The application provides two types of scaffolds: A 50/50 option that highlights two of the incorrect answers in a multiple-choice question and a hint option that provides a prompt that draws students' attention to key parts of the question that might help them to answer it. Figure 4 shows one of the questions in the application with a couple of buttons for the scaffolding mechanism.
Details are in the caption following the image
Screenshot of a multiple-choice question with two buttons (50/50 and hint) for the scaffolding mechanism
  • Feedback: The application provides feedback whether the answer is correct or not (corrective feedback). This type of feedback is shown for each question that students answer in the application. Moreover, the application provides recommendations about the type of question that students need to practice more. This is an integral part of the formative assessment process in which the results that students obtain in the application are used to recommend the content in which students need to practice more.
  • Monitoring module: The mobile application was designed to automatically collect some information about how students interact with the application. It records information about the amount of time that students spend on each question (time on-task), and all the interaction (clickstream) that students do within the application, including the navigation, the correct answers, the type of question that students practice more, the number of times students use the scaffolding mechanisms, as well as the exact moment in which the scaffolding was used. The information collected by the monitoring module is also used to complement the self-reported measures used in this study in an effort to have a more complete and accurate view of how the interaction of students with the mobile-based assessment application might affect their learning outcomes.
  • Appraising: This layer refers to the reports or statistics generated from the data collected from the tests completed by students. In our system, the mobile application shows a ranking in which students can see the number of correct questions answered in the application and a comparison with their peers (see Figure 5). This is another form of feedback implemented in the application. Teachers can also see the amount of time that students have used the application to determine if students need more practice in the application for any specific type of question.
Details are in the caption following the image
Screenshot of the students' ranking with the number of practice tests successfully completed

4.3 Formative assessment activity with the application

As part of the EFL course, students will be invited to participate in a formative assessment activity designed together with EFL teachers in which students have to use the mobile-based assessment application for a period of 4 weeks in preparation for the Cambridge KET exam. Due to the COVID-19 pandemic, students are going to use this application as a support for their learning process at home. This activity was in line with the curriculum and the learning objectives. The application provided corrective feedback on the questions so that students could be aware of the topics they needed to improve. The scaffolding mechanism helped them complete the assessment activity as described in Section 4.2. In this study, the students' learning outcomes in the mobile-assessment application were considered for identifying the factors that might influence them during their use of the application as part of the formative assessment activity. We did not consider other learning outcomes that come from the course exams or the results of the Cambridge Exam.

4.4 Instruments

Data will be collected from two sources: an automatic monitoring mechanism integrated into the mobile-based assessment application and a self-reported instrument. The self-reported instrument consists of the following five scales adapted from the MBAAM (Nikou & Economides, 2017b): Perceived ease of use, perceived usefulness, behavioural intention to use, feedback, and user interface to collect information about the factors that influence the acceptance of mobile-based assessment systems and determine how the acceptance influence the student learning performance. To collect data about motivation, the Student Opinion Scale questionnaire (Thelk et al., 2009) is adapted, which is a well-known questionnaire for collecting data about student motivation in low-stakes tests. Both scales from this questionnaire were used: Importance and Reported Effort.

According to Looi et al. (2015), there are many studies in mobile learning research that have mainly used only self-reported measures. However, self-report measures have some disadvantages. For instance, some instruments do not provide the level of detail needed by the researcher to interpret the results. In other cases, some experiences are subconscious and cannot be fully expressed by people using a questionnaire (Barker et al., 2002). Thus, this study is different from previous research because in this study, we are going to supplement the self-reported measures with an automatic non-intrusive monitoring module developed in the mobile learning application to automatically collect data about how students interact with the application.

4.5 Research procedure

The educational intervention is planned for 4 weeks. Students are going to use the mobile-based assessment application for that period of time. The intervention will not be cross-sectional in order to reduce the novelty of the technology effect or Hawthorne effect (Looi et al., 2009) in the results of this study. By following the recommendations defined by Looi et al. (2015), we integrated the mobile-based assessment application to the curriculum and aligned the use of the application with the learning objectives of the EFL courses. The research procedure is depicted in Figure 6.

Details are in the caption following the image
Research procedure

4.6 Isolating the assessment system and the learning intervention

In this research, the mobile-based assessment application is used to support a formative assessment activity (intervention) in EFL courses. The formative assessment activity implies using the application for a period of 4 weeks, as described in Section 4.3. However, the data collected for validating the structural model is gathered using the self-reported instrument in which the questions refer to the students' experience using the application only and are not related to any other activity in the formative assessment activity or any other experience in the EFL course. In that regard, the information collected to validate the structural model is not affected or biased by the learning intervention itself and allows an accurate result of the predictors of learning performance when using the mobile-based assessment application. Moreover, the monitoring module integrated into the mobile-based assessment application collects information about students' use of the mobile-based assessment application and does not collect any other information about the formative assessment activity.

The hypotheses that can be validated from the data collected with the monitoring module are H1a, H1b, H2 because the variables of time on-task and scaffolding are automatically measured by the mobile-based assessment application. Moreover, the rest of the hypotheses were validated with data collected from the self-reported instrument.

5 HYPOTHESES TESTING (PROPOSAL)

To validate the hypothesized model and identify the determinants of student performance, the PLS-SEM technique will be applied using the SmartPLS software package (Ringle et al., 2015). PLS-SEM is appropriate to explain the variance in the dependent variables due to the influence of independent variables (Hair et al., 2017). Before applying the PLS-SEM technique, we will remove participants with incomplete data or participants who were not able to complete the 4 weeks of the educational intervention in this study.

Following the recommendations to evaluate structural models under the PLS-SEM method (Hair et al., 2017), the evaluation of the structural model proposed is divided into three steps: the evaluation of the formative measurement model (see Section 5.1), the evaluation of the structural model (see Section 5.2) and the evaluation of the model's predictive relevance (see Section 5.3).

5.1 Evaluation of the formative measurement model

The formative measurement model consists of the following exogenous variables with its corresponding indicators (see Figure 1): Scaffolding, time on-task, effort, importance, user interface, feedback, perceived ease of use, perceived usefulness, behavioural intention to use. To evaluate this measurement model, the redundancy analysis will be applied for evaluating the convergent validity, analysis of collinearity (VIF), and analysis of the outer loadings (Hair et al., 2017). The recommended thresholds that will be used to evaluate the measurement model according to Hair et al. (2017) are outer loadings (>0.5), Convergent validity (>0.7), Collinearity (VIF <5), Composite Reliability (>0.7), and Average Variance Extracted (AVE; >0.5).

5.2 Evaluation of the structural model

To evaluate the structural model, the bootstrapping method will be applied in SmartPLS as recommended by Hair et al. (2017). The evaluation of the structural model will provide the amount of variance explained in the student learning performance by the variables in the model. The values of R2, R2 adjusted, total effects, and indirect effects of the structural model will be calculated and provided in this section. Moreover, the effect size (f2) will be calculated in SmartPLS with the p-values and t-values for each path in the structural model.

5.3 Predictive relevance of the structural model

To evaluate the predictive relevance the Stone-Geisser's Q2 measure will be used as recommended by Hair et al. (2017). This measure gives an idea of the extent to which the model is able to predict new values. To obtain the Stone-Geisser's Q2 measure, the Blindfolding method with an omission distance of seven and using the cross-validated redundancy approach can be applied in SmartPLS. The purpose of this evaluation is to obtain an adequate predictive power and effect sizes in the model proposed.

CONFLICT OF INTEREST

The authors declare no conflict of interest.

PEER REVIEW

The peer review history for this article is available at https://publons.com/publon/10.1111/jcal.12649.

DATA AVAILABILITY STATEMENT

N/a.

    The full text of this article hosted at iucr.org is unavailable due to technical difficulties.