Educational Research and Reviews

  • Abbreviation: Educ. Res. Rev.
  • Language: English
  • ISSN: 1990-3839
  • DOI: 10.5897/ERR
  • Start Year: 2006
  • Published Articles: 2026

Full Length Research Paper

Attitude scale towards web-based examination system (MOODLE) - Validity and reliability study

Basaran Bulent
  • Basaran Bulent
  • Computer Education and Instructional Technology, Dicle University, Turkey.
  • Google Scholar
Yalman Murat*
  • Yalman Murat*
  • Computer Education and Instructional Technology, Dicle University, Turkey.
  • Google Scholar
Gonen Selahattin
  • Gonen Selahattin
  • Department of Physics Education, Dicle University, Turkey.
  • Google Scholar


  •  Received: 28 June 2016
  •  Accepted: 18 August 2016
  •  Published: 10 September 2016

 ABSTRACT

Today, the spread of Internet use has accelerated the development of educational technologies and increased the quality of education by encouraging teachers’ cooperation and participation. As a result, examinations executed via the Internet have become common, and a number of universities have started using distant education management system. Eventually, today, more people have a chance to take education. Measurement and evaluation applications carried out via the Internet are now quite important for education. In the present study, a valid and reliable attitude scale was developed to measure the attitudes of students doing the Distant Education Theology Undergraduate Education Program towards a web-based examination system (MOODLE).The study group was made up of 1287 3rd and 4th grade students registered in the Distant Education Application and Research Center. In the study, exploratory factor analysis (EFA) was conducted to determine the factor structure of the scale; confirmatory factor analysis (CFA) was conducted to test its construct validity; and other validity analyses such as exploratory factor analysis, uni-dimensional factor analysis and two-dimensional factor analysis were applied. The research data were analyzed in computer with the package softwares of SPSS 18.0 and Lisrel 8.51.

Key words: Attitude, moodle, education technology.


 INTRODUCTION

In the past decade, rapid developments in information and communication technologies (ICT) have led to the developments in the field of education and increased the quality of education besides contributing to learning experiences (Yalman and Tunga, 2014). This rapid development of ICT has enriched teaching and learning experiences and led to a better-quality education (Tella, 2011; Maldonado et al., 2011). Due to the rapid growth of the Internet, e-learning has become an alternative that has facilitated students’ learning (Wang Tzu-Hua, 2008). When compared with the traditional learning, it is seen that e-learning provides students with more sources and allows them to gain more satisfactory learning experiences via instructional activities (Saulnier et al., 2008). Thanks to e-learning, students can carry out learning activities by determining the time and  their  pace of learning the subjects. In other words, with the help of e-learning materials, students have the chance to evaluate their own advances. In face-to-face higher education, evaluation methods could be central as a basic component of effective learning and can define such methods in the learning process as measurement of students’ achievements and their pace of learning (Gikandi et al., 2011). Mid-term exams, end-of-term exams and quizzes may be given as examples of evaluation. In general, examination is one of the main methods used to confirm the results of students’ learning. Quizzes can be conducted to produce better instructional materials and for faculty members to evaluate students’ learning experiences. In this way, the quality of students’ academic achievements and their learning experiences can be increased (Miguel et al., 2016; Jordan, 2012; Levy and Ramim, 2007).

It is known that different methods used to increase students’ academic achievements positively contribute to education levels. Probably, students’ attitudes constitute one of the most important processes that increase academic achievement (Gül et al., 2015). Attitude can be named as a mental process in which individuals determine their own behaviors in certain situations (Gagne, 1985). According to Smith (1968), attitude can be defined as a tendency that forms individuals’ thoughts, emotions and behaviors regarding a regular psychological object. An attitude and a learned fact shape individuals’ behaviors, can lead to a bias in the process of decision making. Attitudes occur as a result of the learning process and experiences (TavÅŸancıl, 2006). Online evaluations can be used in the evaluation of academic achievement. In addition, it contributes to meaningful learning and delivers the concepts or the learned information to the student as feedback. In traditional class applications, when teachers ask students a question, students are provided with a little chance of responding to the question. This makes it difficult for other students to see whether they have understood the subject or not. When questions are directed to students in distant education management system, the system can instantly provide online feedback and students thus, become more successful in learning when compared to the traditional system (Robles and Braathen, 2002). Online evaluation methods have a number of advantages mentioned above, yet there are limited areas for their application in higher education (Wen and Tsai, 2006). On the other hand, distance education at universities and private education institutions is still popular since it allows reaching a large population of students accommodating in different geographical regions.

The number of distant education software is gradually increasing, and more students can now access these software. Preference of online test to evaluate students’ performances in educational processes will allow conducting    both    test    applications     and     reporting procedures more rapidly. Attitudes are undoubtedly regarded as a good determiner of academic achievement. It is seen that technology has been in use to a great extent in educational studies in recent years. However, it is obvious that the number of studies conducted to reveal students’ attitudes towards online evaluations is quite limited (Dermo, 2009). In general, technology-aided studies were measured in relation to technology and computer, and their academic achievements were predicted by these variables. In addition, although, there are positive attitudes towards computers, students are likely to demonstrate negative attitudes towards online evaluation (Bindak and Çelik, 2006; Ergün, 2002).

 

Theoretical framework

MOODLE has been used as an education platform via the Internet by a number of public institutions and private corporations for years. Many researchers investigated general system features of these web-based learning systems, users’ attitudes towards these systems and their levels of satisfaction with these systems (Coates et al., 2005; Engelbrecht, 2005; Marikar and Jayarathne, 2016; Martínez-Torres et al., 2008; Njenga and Fourie, 2010; Seale and Cooper, 2010). There several reasons for this popularity of MOODLE: it is free of charge to access the system; it is an open-source system; and users can easily solve problems themselves (Gutiérrez et al., 2010; Kakasevski et al., 2008; Limongelli et al., 2011; Xu and Mahenthiran, 2016). This education system, whose courses or curriculum can be designed in line with users’ needs via the web, have such basic features as homework and source sharing, questionnaires and forums as well as a testing system that allows measuring students’ success at the end of the education given. The study aimed at determining the views and attitudes of students taking education via the distance education management system towards the exam system. In this way, it could be easier to see whether the whole system functions well or not.

Recent developments in information and communication technologies have caused a number of published papers to become out of date. In this respect, in web-aided instructional methods, information technologies are shaped in accordance with the renovations since these technologies are influenced by the related developments. In literature, when the papers in the field of web-aided education are examined, it is seen that issues mentioned by other studies were mostly examined via generalization (Brine et al., 2007; Georgouli et al., 2008; Zakaria and Daud, 2008). On the other hand, although, such systems are similar to each other in terms of function, methodology and form, their user interfaces, courses, questionnaires and  exam  systems  could  differ (Brine et al., 2007; Romero et al., 2009). The basic features used here such as presentation of courses, questionnaires and forms are for general use of the system, while the exam system is used to measure and evaluate the success of students. Correct and valid measurement of students’ gains at the end of an education process depends on the features found in the exam module. Students’ evaluation of the exam module to determine its negative or positive aspects is important to reveal the related deficiencies. This study focused just on determining the attitudes towards the exam module used in MOODLE system rather than determining the attitudes towards exam modules used in web-aided education systems.

Evaluation of the success of the web-based learning system of MOODLE, which has millions of users all over the world, will be possible via the related measurements and the results to be obtained from these measurements. When the related literature is examined, it is seen that there are studies which examined attitudes of users towards e-learning environments (Graf et al., 2009; Kakasevski et al., 2008; Sun et al., 2008)as well as those which investigate students’ views about the system (Kao and Tsai, 2009; Richardson, 2009; Sher, 2009; Yassine et al., 2016). Depending on the results of these studies, it could be stated that it will be better to overcome the problems existing or to exist in the system or to improve the popular applications in line with users’ feedbacks.

Measurement and evaluation constitute the basis of education. All the evaluations to determine students’ levels of knowledge about the subjects they have been taught, should include objective and healthy evaluations. In contrast with the exams given in traditional education, those conducted in web-aided education include different features and norms; Questions should be clear and comprehensible; time limitations should focus on solving the questions; and a simple language should be used in the instructions to solve the questions. Here, the purpose is to minimize the probable problems. Students’ evaluations regarding the web-aided exam system at the end of the exams will help overcome related future problems. 


 METHODS

The participants in the study constituted those taking their theology undergraduate education via distance education system using the platform of MOODLE. In the study, a scale was developed to evaluate the students’ attitudes towards the web-based exam system (MOODLE). The data collected in the scale development process were analyzed with “Exploratory Factor Analysis”, “Uni-Dimensional Data Analysis” and “Two-Dimensional Factor Analysis” to examine appropriateness of the values to the fit indices.

 

Sample

The participants in the study were 1300 students who registered in Theology Undergraduate Education Program; they received education with the e-learning management system in a distant education center of a state university. The scale designed for the study was applied via the system which was used to give education. A total of 1287 students responded to the scale conducted online. In this way, almost the whole research group was reached. In the study, as for the demographic backgrounds of the students filling out the scale in the study, 45.53% of them were female, and 54.47% of them were male. The data collected were examined in digital environment. Among the responses to the scale, those with all of its options marked same, those with most of its options marked same and those given to reverse items were examined, and the questionnaire forms which were not responded to correctly or completely were not included in the scope of the study to preserve the objectivity of the study. As a result, a total of 653 questionnaire forms were included in the study. Table 1 below presents the frequencies and percentages of the participants with respect to their gender. According to the demographic backgrounds of the participants, 51.61% of them were females, and 48.39% of them were males (Table 1).

 

 

Attitude scale for the web-based learning process

For the purpose of determining the success, level and quality of education given via the web, several scales have been developed by researchers. Examining such scales in the study (Dermo, 2009; Gül et al., 2015; Bahar, 2014) and considering the standards previously determined (Swedish National Agency for Higher Education, 2008), a trail item pool was formed depending on the related literature, and faculty members expert in the field were asked for their views. The scale was presented to three experts in the field and to two Turkish language experts. The scale made up of 31 items, 23 of which were positive and eight of which were negative, was piloted to see whether it was comprehensible or not. The results of the statistical analysis of the collected data revealed that five items were difficult to comprehend and 10 items considered by the experts to be statistically inappropriate were excluded. As a result, there were 16 items in the finalized scale. Following this, the updated version of the scale was applied again to collect the research data. The scale included four items in the dimension of “System and Usability”, three items in the dimension of “Comprehensibility”, six items in the dimension of “Examination and Features” and three items in the dimension of “Security and Reliability”. The scale was designed as Five-point Likert-type with the choices of 1- I completely disagree, 2- I disagree, 3- I partly agree, 4- I agree and 5- I completely agree. In attitude scales, a five-point rating method ranging from “I completely agree” to “I completely disagree” could be used (Dunn-Rankin, 2004; TavÅŸancıl, 2005). When the related literature is examined, it is seen that scale development phases are as follows (TavÅŸancıl, 2005; Dunn-Rankin, 2004; Devellis, 2003; Karasar, 1995).

1. Forming the item pool M

2. Asking for experts’ view

3. Conducting the pilot application

4. Applying the draft scale to the study group and factor analyses

5. Calculating the reliability of the scale

In order to determine the factor loads predicted for scale development, exploratory factor analysis (EFA) and confirmatory factor analysis (CFA) were conducted. After the results of these two analyses that were obtained, confirmatory factor analysis was used for the model data fit. The fit indices used in the study were Chi-Square fit test, Goodness of Fit Index (GFI), Adjusted Goodness of Fit Index (AGFI), Root Mean Square (RMS) and Root Mean Square Error of Approximation (RMSEA). Table 2 presents the results obtained via the analysis of the data. The value of x2/sd obtained in the study was lower than 3, which demonstrated that the model had an acceptable fit (ÅžimÅŸek, 2007). The value in question was higher than three. As the value of x2/sd is sensitive to the size of the sample, it should be interpreted with other fit indices (Jöreskog and Sörbom, 1999). According to the model data fit, the value of 0.94 GFI was higher than 0.09; the value of 0.91 AGFI was higher than 0.09; the value of 0.064 RMSEA was lower than 0.08; and the value of 0.038 RMR was lower than 0.05.

 


 FINDINGS

Exploratory factor analysis

The scale developed was applied to 1288 individuals, yet the analysis was carried out with 653 participants due to the fact that some of the questionnaires were not filled completely or some of them included items all with the same options marked. The Kaiser-Mayer-Olkin (KMO) value of the scale was 0.877, and Barlett’s test was found significant (p < 0.01). The scale included a total of 16 items and four factors. In line with the expert view, no change was done in the scale. Table 3 presents the results of the exploratory factor analysis. 

 

 

Uni-dimensional confirmatory factor analysis (CFA)

Figure 1 presents the results of the confirmatory factor analysis conducted to determine the fit between the factors and items in the scale. According to the results of the confirmatory factor analysis, the Chi-Square  value  of x2(98,N=685)=7663,63 calculated for the model-data fit was significant (p < 0.000). The fit statistics values calculated using the Lisrel GFI = 0.94, AGFI = 0.91, CFI = 0.97, NNFI = 0.96 software were as follows: RMSEA = 0.064, RMR = 0.038, and = 0.97 dir. Since these values were in appropriate ranges, there was no need for any modification in the scale.

IFI

 

 

Two-dimensional confirmatory factor analysis (CFA)

The attitude scale for the web-based learning process with its 16 items and four factors was tested with two-dimensional confirmatory factor analysis (CFA). According to EFA, the items for the fact of System and usability had standard solutions of 0.64, 0.66, 0.58 and 0.51; those for the fact of Examination and Features had standard solutions of 0.49, 0.49, 0.54, 0.56, 0.54 and 0.53; those for the factor of Comprehensibility had standard solutions of 0.75, 0.71 and 0.90; and those for the factor of Security and Reliability had standard solutions of 0.53, 0.54 and 0.49. As can be seen, all the standard solutions were found to be higher than 0.45. As a result of CFA, the fit statistics values calculated as x2/sd = 3.36 were as follows: RMSEA = 0.064, RMR = 0.038, GFI = 0.94, AGFI = 0.91, CFI = 0.97, NNFI = 0.96 and IFI = 0.97. All the fit indices obtained were found to demonstrate an acceptable fit according to Schermelleh-Engel, Moosbrugger and Müller (2003) (Figure 2). It was seen that the values obtained regarding the items and the whole scale were in acceptable ranges for the applicability of the scale.

 

 

Results of reliability analysis

Table 4 presents the reliability coefficients for each factor in the scale. The Cronbach Alpha value for the whole scale was calculated as 0.873. The Cronbach Alpha values were calculated as 0.834 for the sub-factor of “System and Availability”, as 0.793 for the factor “Comprehensibility”, as 0.761 for the factor of “Examination and Features” and as 0.740 for the factor of “Safety and Reliability”. The Cronbach’s alpha values obtained in relation to the sub-factors demonstrate that the scale developed was valid (Brownlow, 2004).

 


 DISCUSSIONS

Depending on the related need in literature, the present study aimed at developing a valid and reliable attitude scale to evaluate the attitudes of distant education theology undergraduate students towards the Web-Based Examination System (Appendix). For this purpose, in line  with  the  scale  development  phases  reported  in   related literature (Devellis, 2003; Dunn-Rankin, 2004; Karasar, 1995; TavÅŸancıl, 2005), an item pool was formed; experts’ views were taken regarding the items; and the trial item pool was piloted. Following the pilot application, the draft scale was applied to the study group, and the data were examined in terms of the distribution of the scale scores obtained before conducting the Exploratory Factor Analysis. After the distribution was found to be at the desired level, to determine the factor structure of the scale, basic components analysis was preferred as the factorization method, and the maximum varimax technique, one of the vertical rotation method, was preferred considering and significance. In the study, Exploratory Factor Analysis was conducted sequentially and gradually on the data several times. For the statistical evaluation of the data collected during the application, five items lacking unity of meaning and ten items considered statistically to be inappropriate were excluded from the scope of the study.

After the scale items were updated, the scale was applied again to collect the research data. The final version of the scale, which was initially made up of 31 items, included 16 items. In addition, factor analysis applied again to the final version of the scale. According to the results of the exploratory factor analysis, there were four factors with an Eigen value higher than 1. It was found that the first factor (System and Availability) contributed to the common variance with a rate of 17,154%; the second factor (Comprehensibility) with a rate of 13,834%; the third factor (Examination and Features) with a rate of 16,496%; and the fourth factor (Safety and Reliability) contributed to the common variance with a rate of 12,868%. The contribution of the four factors to the total variance was calculated as 60,353%. Also, the Kaiser-Mayer-Olkin (KMO) value of the scale was calculated as 0.877, and Barlett’s test was found significant (p < 0.01). In order to evaluate the validity of the two-factor structure obtained as a result of the exploratory factor analysis, confirmatory factor analysis (CFA) was conducted. When the CFA fit values were examined, the Chi-Square value of x2(98,N=685) = 51.38 calculated for the model-data fit was found significant (p <0.000). The fit statistics values calculated as a result of the analysis carried out with Lisrel software were as follows: RMSEA = 0.064, RMR = 0.038, GFI = 0.94, AGFI = 0.91, CFI = 0.97, NNFI = 0.96 and IFI = 0.97. Since the values obtained were in appropriate ranges, no modification was done in the scale.

The web-based learning attitude scale made up of 16 items and four factors was also tested with the two-dimensional confirmatory factor analysis (CFA). The fit statistics values found to be x2/sd = 3.36 as a result of CFA were as follows: RMSEA = 0.064, RMR = 0.038, GFI = 0.94, AGFI = 0.91, CFI = 0.97, NNFI = 0.96 and IFI = 0.97. According to Schermelleh-Engel, Moosbrugger and Müller (2003), all the fit indices obtained were found to demonstrate acceptable fit. In addition, the Cronbach Alpha values were calculated as, 834; for the factor of “System and Availability”, as, 793; for the factor of “Comprehensibility”, as, 761; for the factor of “Examination and Features” and as, 740 for the factor of “Safety and Reliability”. The scale could be used as a valid and reliable data collection tool in studies to be conducted not only to determine the attitudes of students from the Department of Theology Undergraduate Education Program towards web-based distant education as well as towards related exam practices but also to examine the factors influential on their attitudes. 


 CONCLUSION

Rapid changes and developments in computer and Internet technologies lead to the development and spread of e-learning management systems. In this process of changes, e-evaluation has gradually gained more importance and become a significant part of this transformation. Evaluation on paper-pen basis is quite an inefficient method of evaluating students’ success and making education decisions (Sırakaya et al., 2014). Although, the reliability and objectivity of exams conducted via e-learning platforms are examined by researchers and experts, these exams have a number of advantages when compared to those conducted with traditional methods. The most prominent advantages of e-evaluation include saving time and cost, gathering the responses to the questions in computer environment, providing appropriate and rapid feedback, allowing flexibility, increasing the reliability by minimizing the mistakes made by human (Struyven, Dochy and Janssens 2002; Angus and Watson 2009), decreasing faculty members’ involvement (Anderson et al., 2005) and obtaining the results rapidly (Kuhtman, 2004). In addition, exams conducted in computer environment, in contrast with paper-pen exams, allow enriching the presentation of information via integration of multimedia elements (Liu et al., 2001). Obviously, electronic evaluations are beneficial. Faculty members can obtain rapid results, thanks to e-evaluation, and use of computers that will make education better. This will allow transition from a traditional education environment to the one that makes students more active and contributes to their learning. In recent years, the traditional measurement and evaluation methods in the field of education have been replaced by e-evalaution. Thanks to this, the field of measurement and evaluation has gained a new dimension. As an alternative, faculty members try to spread the use of the Internet and computer in the field of measurement and evaluation (Özmen, 2006; Lawrenz et al., 2001).


 CONFLICT OF INTERESTS

The authors have not declared any conflict of interests.



 REFERENCES

Anderson HM, Cain J, Bird E (2005). Online course evaluations: Review of literature and a pilot study. Am. J. Pharmaceutical Educ. 69(1):34-43.
Crossref

 

Angus SD, Watson J (2009) Does regular online testing enhance student learning in the numerical sciences? Robust evidence from a large data set. Br. J. Educ. Technol. 40:255-272.
Crossref

 

Bahar M (2014).Üniversite ÖÄŸrencileri için E-DeÄŸerlendirme Tutum ÖlçeÄŸinin GeliÅŸtirilmesi.e-International J. Educ. Res. 5(2):38-53.

 

Bindak R, Çelik HC (2006). ÖÄŸretmenler için Bilgisayar Tutum ÖlçeÄŸinin Güvenirlik Geçerlik Çalışması, Eurasian J. Educ. Res. 22:38-47.

 

Brine J, Wilson I, Roy D (2007). Using moodle and other software tools in EFL courses in a Japanese IT University. Paper presented at the Computer and Information Technology, 2007.
Crossref

 

Brownlow C (2004). SPSS explained. London: Routledge.

 

Coates H, James R, Baldwin G (2005). A critical examination of the effects of learning management systems on university teaching and learning. Tertiary Educ. Manage. 11:19-36.
Crossref

 

Dermo J (2009). E-Assessment and the student learning experience: A survey of student perceptions of e-assessment. Br. J. Educ. Technol. 40(2):203-214.
Crossref

 

Devellis R (2003). Scale Development Theoryand Applications. London: SAGE.

 

Dunn-Rankin P (2004). Scaling Methods. London: Routledge.

 

Engelbrecht E (2005). Adapting to changing expectations: Post-graduate students' experience of an e-learning tax program. Comput. Educ. 45(2):217-229.
Crossref

 

Ergün E (2002). Üniversite ÖÄŸrencilerinin Bilgisayar Destekli Ölçmeden Elde Ettikleri Basarının Kalem-Kâğıt Testi BaÅŸarası, Bilgisayar Kaygısı ve Bilgisayar Tecrübeleri Açısından İncelenmesi. Yayımlanmamış yüksek lisans tezi. Anadolu Üniversitesi EÄŸitim Bilimleri Enstitüsü, EskiÅŸehir.

 

Gagne RMT (1985). The Conditions of Learning and Theory of fnsmtctlon (4th ed.). Holt, Rinehart & Winston, New York.

 

Georgouli K, Skalkidis I, Guerreiro P (2008). A Framework for Adopting LMS to Introduce e-Learning in a Traditional Course. Educ. Technol. Society, 11(2):227-240.

 

Gikandi JW, Morrow D, Davis NE (2011). Online formative assessment in higher education: A review of the literatüre.Comput. Educ. 57:2333-2351.
Crossref

 

Graf S, Liu TC, Chen NS, Yang SJ (2009). Learning styles and cognitive traits–Their relationship and its benefits in web-based educational systems. Comput. Hum. Behav. 25(6):1280-1289.
Crossref

 

Gutiérrez E, Trenas MA, Ramos J, Corbera F, Romero S (2010). A new Moodle module supporting automatic verification of VHDL-based assignments. Comput. Educ. 54(2):562-577.
Crossref

 

Gül E, Çokluk Ö, Gül ÇD (2015) . Development of an Attitudes Scale toward Online Assessment. Procedia – Soc. Behav. Sci. 174:529-536.
Crossref

 

Jordan S (2012). Student engagement with assessment and feedback: some lessons from short-answer free-text e-assessment questions. Comput. Educ. 58(2):818-834.
Crossref

 

Jöreskog KG, Sörbom D (1999). Lisrel 8.30 and Prelis 2.30. In: Scientific SoftwareInternational.

 

Kakasevski G, Mihajlov M, Arsenovski S, Chungurski S (2008). Evaluating usability in learning management system Moodle. Paper presented at the Information Technology Interfaces, 2008. ITI 2008. 30th International Conference on.
Crossref

 

Karasar N (1995). Bilimsel araÅŸtırma yöntemi, kavramlar, ilkeler. Ankara: 3A AraÅŸtırma EÄŸitim Danışmanlık Ltd.

 

Kao CP, Tsai CC (2009). Teachers' attitudes toward web-based professional development, with relation to Internet self-efficacy and beliefs about web-based learning. Comput. Educ. 53(1):66-73.
Crossref

 

Kuhtman M (2004). Review of online student ratings of instruction. College and University J. 80(1):64-67.

 

Lawrenz F, Huffman D, Welch W (2001). The science achievement of various subgroups on alternative assessment formats. Sci. Educ. 85:279-290.
Crossref

 

Levy Y, Ramim M (2007). A theoretical approach for biometrics authentication of e-exams. Nova Southeastern University, USA, pp. 93-101.

 

Limongelli C, Sciarrone F, Vaste G (2011). Personalized e-learning in Moodle: the Moodle_LS System. J. e-Learn. Knowledge Soc. 7(1).

 

Liu M, Papathanasiou E, Yung-Wei H (2001) Exploring the use of multimedia examination formats in undergraduate teaching: results from the fielding testing. Comput. Hum. Behav. 17:225-248.
Crossref

 

Maldonado UPT, Khan GF, Moon J, Rho JJ (2011). 'E-learning motivation and educational portal acceptance in developing countries.' Online Infor. Rev. 35(1):66–85. doi:10.1108/14684521111113597
Crossref

 

Marikar FM, Jayarathne N (2016). Effectiveness of MOODLE in Education System in Sri Lankan University. Int. J. Modern Educ. Comput. Sci. 8(2):54.
Crossref

 

Martínez-Torres MR, Toral Marín S, Garcia FB, Vazquez SG, Oliva MA, Torres T (2008). A technological acceptance of e-learning tools used in practical and laboratory teaching, according to the European higher education area 1. Behav. Inform. Technol. 27(6):495-505.
Crossref

 

Miguel J, Caballé S, Xhafa F, Prieto J, amp Barolli L (2016). A methodological approach for trustworthiness assessment and prediction in mobile online collaborative learning. Comput. Standards Interfaces 44:122-136.
Crossref

 

Njenga JK, Fourie LCH (2010). The myths about e‐learning in higher education. Br. J. Educ. Technol. 41(2):199-212.
Crossref

 

Özmen H (2006). E-Sınav Sistemi. Akademik BiliÅŸim, Pamukkale University, Denizli.

 

Richardson JT (2009). The attainment and experiences of disabled students in distance education. Distance Educ. 30(1):87-102.
Crossref

 

Robles B, Braathen S (2002). Online assessment techniques. Delta Pi Epsilon Journal 44(1).

 

Romero C, González P, Ventura S, Del Jesús MJ, Herrera F (2009). Evolutionary algorithms for subgroup discovery in e-learning: A practical application using Moodle data. Expert Systems Applications 36(2):1632-1644.
Crossref

 

Saulnier BM, Landry JP, Longenecker Jr HE, Wagner TA (2008). "From Teaching to Learning: Learner-Centered Teaching and Assessment in Information Systems Education," J. Infor. Syst. Educ. 19(2):169-174, Summer 2008.

 

Seale J, Cooper M (2010). E-learning and accessibility: An exploration of the potential role of generic pedagogical tools. Comput. Educ. 54(4):1107-1116.
Crossref

 

Schermelleh-Engel K, Moosbrugger H, Müller H (2003). Evaluating the fit of structural equation models: Tests of significance and descriptive goodness-of-fit measures. Methods Psychological Res. Online 8(2):23-74.

 

Sher A (2009). Assessing the relationship of student-instructor and student-student interaction to student learning and satisfaction in web-based online learning environment. J. Interactive Online Learn. 8(2):102-120.

 

Sırakaya M, Sırakaya DA, Çakmak EK (2014). Uzaktan EÄŸitim ÖÄŸrencilerinin Çevrimiçi Sınava Yönelik Tutum Düzeylerinin İncelenmesi. Kastamonu EÄŸitim Dergisi 23(1):87-104.

 

Struyven K, Dochy F, Janssens S (2002). Students&#39; perceptions about assessment in higher education: a review. Paper presented at the Joint Northumbria/ Earli SIG Assessment and Evaluation Conference: Learning communities and assessment cultures, University of Northumbria at Newcastle, Stödberg, U. (2011). A research review of e-assessment Assessment &amp; Evaluation in Higher Education, 2011, 1–14, 
Crossref

 

Sun PC, Tsai RJ, Finger G, Chen YY, Yeh D (2008). What drives a successful e-Learning? An empirical investigation of the critical factors influencing learner satisfaction. Comput. Educ. 50(4):1183-1202.
Crossref

 

ÅžimÅŸek ÖF (2007). Yapısal eÅŸitlik modellemesine giriÅŸ: Temel ilkeler ve LISRELuygulamaları. Ankara: Ekinoks Yayınları.

 

TavÅŸancıl E (2005). Tutumların ölçülmesi ve spss ile veri analizi. Ankara: Nobel Yayınları.

 

TavÅŸancıl E (2006), Tutumların Ölçülmesi ve SPSS ile Veri Analizi, Ankara, Nobel Yayın Dağıtım.

 

Tella A (2011). Reliability and factor analysis of a blackboard course management system success: A scale development and validation in an educational context. J. Inform. Technol. Educ. 10(1):55-80.

 

Xu H, Mahenthiran S (2016). Factors that Influence Online Learning Assessment and Satisfaction: Using Moodle as a Learning Management System. Int. Bus. Res. 9(2):1.
Crossref

 

Wang Tzu-Hua (2008). Web-based quiz-game-like formative assessment: Development and evaluation. Comput. Educ. 51:1247-1263.
Crossref

 

Wen ML, Tsai CC (2006). University students' perceptions of and attitudes toward (online) peer assessment. Higher Educ. 51:27-44.
Crossref

 

Yalman M, Tunga MA (2014). Examining the attitudes of students from state and foundation universities in Turkey towards the computer and www (world wide web). Egitim ve Bilim 39(173).

 

Yassine S, Kadry S, Sicilia MA (2016). A framework for learning analytics in moodle for assessing course outcomes. Paper presented at the 2016 IEEE Global Engineering Education Conference (EDUCON).
Crossref

 

Zakaria E, Daud M (2008). Using MOODLE in a Mathematical Methods Course: Pre-Service Teachers' Perspective. Int. J. Learn. 15(4).

 




          */?>