<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Intelligently Raising Academic Performance Alerts</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Dimitris</forename><surname>Kalles</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Hellenic Open University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Christos</forename><surname>Pierrakeas</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Hellenic Open University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Michalis</forename><surname>Xenos</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Hellenic Open University</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Intelligently Raising Academic Performance Alerts</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">C56231A94508D2E8B9510B9579D103B2</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T21:06+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We use decision trees and genetic algorithms to analyze the academic performance of students and the homogeneity of tutoring teams in the undergraduate program on Informatics at the Hellenic Open University (HOU). Based on the accuracy of the generated rules, we examine the applicability of the techniques at large and reflect on how one can deploy such techniques in academic performance alert systems.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>Student success is a natural performance indicator in universities. However, if that success is used as a criterion for tutor assessment (and subsequent possible contract renewal), and if students must evaluate their own teachers, then tutors may tend to lax their standards. This paper is about dealing with this issue in the context of the Hellenic Open University (HOU); we focus on the undergraduate Informatics program (about 2,500 students). We ask whether we can detect regularities in distance tutoring, then, we try to associate them with measures of students' success in an objective way and, subsequently, reflect on how to effectively disseminate this information to all interested parties.</p><p>The measurement strategy we have developed to-date in HOU has been progressively refined to deal with two closely linked problems: that of predicting student success in the final exams and that of analyzing whether some specific tutoring practices have any effect on the performance of students. Each problem gives rise to the emergence of a different type of user model. A student model allows us, in principle, to explain and maybe predict why some students fail in the exams while others succeed. A tutor model allows us to infer the extent to which a group of tutors diffuses its collective capacity effectively into the student population they advise. However, both types of models can be subsequently interpreted in terms of the effectiveness of the educational system that the university implements.</p><p>The rest of this paper is organised in five sections. The next section presents the educational background. Section 3 then reviews the fundamental features of the AI techniques that we have used. Following that we report the experimental results for the undergraduate programme that we have analysed, as well as a short evaluation of the individual module results that seem to signify an interesting deviation. Section 5 presents a discussion from the point of view of how one can generalise our approach as well as how one can substitute other intelligent techniques for data analysis; finally we conclude and describe directions for future development.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">THE EDUCATIONAL BACKGROUND</head><p>A module is the basic educational unit at HOU. It runs for about ten months and is the equivalent of about 3-4 conventional university semester courses. A student may register with up to three modules per year. For each module, a student is expected to attend five plenary class meetings throughout the academic year. A typical class contains about thirty students and is assigned to a tutor (tutors of classes of the same module collaborate on various course aspects). Class face-to-face meetings are about four hours long and are structured along tutor presentations, group-work and review of homework. Furthermore, each student must turn in some written assignments (typically four or six), which contribute towards the final grade, before sitting a written exam. That exam is delivered in two stages: you only need sit the second if you fail or miss the first.</p><p>Students fail a module and may not sit the written exam if they do not achieve a pass grade in the assignments they turn in; these students must repeat that module afresh. A student who only fails the written exam may sit it on the following academic year (without having to turn in assignments); such "virtual" students are also assigned to student groups but the tutor is only responsible for marking their exam papers.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">GENETIC ALGORITHMS AND DECISION TREES FOR PREDICTION</head><p>In our work we have relied on decision trees to produce performance models. Decision trees can be considered as rule representations that, besides being accurate, can produce comprehensible output, which can be also evaluated from a qualitative point of view <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>. In a decision tree nodes contain test attributes and leaves contain class descriptors.</p><p>A decision tree for the (student) exam success analysis problem could look like the one in Figure <ref type="figure" target="#fig_0">1</ref> and tells us that a mediocre grade at the second assignment (root) is an indicator of possible failure (left branch) at the exams, whereas a non-mediocre grade refers the alert to the fourth (last) assignment.</p><p>Decision trees are usually produced by analyzing the structure of examples (training instances), which are given in a tabular form. An excerpt of a training set that could have produced such a tree is shown in Table <ref type="table" target="#tab_0">1</ref>. Note that the three examples shown are consistent with the decision tree. As this may not always be the case, there rises the need to measure accuracy, even on the training set, in order to compare the quality of two decision trees which offer competing explanations for the same data set. Note that the sample decision tree does not utilize data neither on the first nor the third assignments, but such data is shown in the associated table. Such dimensionality reduction information is typical of why decision trees are useful; if we consistently derive trees on some problem that seem to not use some data column, we feel quite safe to not collect measurements for that data column. Of course, simple correlation could also deliver such information, however it is the visual representation advantages of decision trees that have rendered them as very popular data analysis tools. Analyzing the performance of high-risk students is a goal towards achieving tutoring excellence. It is, thus, reasonable to assert that predicting a student's performance can enable a tutor to take early remedial measures by providing more focused coaching, especially in issues such as priority setting and time management.</p><p>Initial experimentation at HOU <ref type="bibr" target="#b3">[4]</ref> consisted of using several machine learning techniques to predict student performance with reference to the final examination. The scope of the experimentation was to investigate the effectiveness and efficiency of machine learning techniques in such a context. The WEKA toolkit <ref type="bibr" target="#b4">[5]</ref> was used because it supports a diverse collection of techniques. The key result was that learning algorithms could enable tutors to predict student performance with satisfying accuracy long before final examination. The key finding that lead to that result was that success in the initial written assignments is a strong indicator of success in the examination. Furthermore, our tutoring experience corroborates that finding.</p><p>We then employed the GATREE system <ref type="bibr" target="#b5">[6]</ref> as the tool of choice for our experiments, to progressively set and test hypotheses of increasing complexity based on the data sets that were available from the university registry. The formation and development of these tests is the core content of this chapter and is presented and discussed in detail in the following sections. GATREE is a decision tree builder that employs genetic algorithms to evolve populations of decision trees; it was eventually used because it produces short comprehensible trees.</p><p>Of course, GATREE was first used <ref type="bibr" target="#b2">[3]</ref> to confirm the qualitative validity of the original findings experiments <ref type="bibr" target="#b3">[4]</ref>, also serving as result replication, before advancing to more elaborate experiments <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9]</ref>. GATREE <ref type="bibr" target="#b5">[6]</ref> evolves populations of trees according to a fitness function that allows for fine-tuning decision tree size vs. accuracy on the training set. At each generation, a certain population of decision trees is generated and sorted according to fitness. Based on that ordering, certain genetic operators are performed on some members of the population to produce a new population. For example, a mutation may modify the test attribute at a node or the class label at a leaf, while a cross-over may exchange parts between decision trees.</p><p>The fitness function is fitness i =Correct i 2 *x/(size i 2 +x), for tree i. The first part of the product is the actual number of training instances that i classifies correctly. The second part of the product (the size factor) includes a factor x which regulates the relative contribution of the tree size into the overall fitness; thus, the payoff is greater for smaller trees When using GATREE, we used the default settings for the genetic algorithm operations and set cross-over probability at 0.99 and mutation probability at 0.01. Moreover, all but the simplest experiments (explicitly so identified in the following sections) were carried out using 10-fold cross-validation, on which all averages are based (i.e. one-tenth of the training set was reserved for testing purposes and the model was built by training on the remaining nine-tenths; furthermore, ten such stages were carried out by rotating the testing one-tenth.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">DATA ANALYSIS AT A PROGRAMME LEVEL</head><p>Before advancing, we first review some aggregate statistics of the undergraduate informatics programme at HOU. First, Table <ref type="table" target="#tab_1">2</ref> presents the success rates for the modules that we have analysed. Next, Table <ref type="table" target="#tab_2">3</ref> presents the enrolment numbers for these modules. Note that, as we advance from junior to senior years, the overall enrolment is dramatically reduced and the success rates increase. The above statistics are all drawn from the university registry and none is subject to any further processing. However, all results presented from now on, refer to experiments carried out totally using the GATREE system, with the occasional help of some post-processing automation scripts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Detecting a shift in exam grades</head><p>There is a straightforward way to attempt to answer this question. One can build a model that attempts to answer the success question for the first stage of the final exam. Then, one can build a model that attempts to answer the success question for the overall student grade. A gross disparity in these numbers should be indicative of an issue that merits investigation.</p><p>The simplest data to consider as input for this problem consists of exercise and exam grades, as in Table <ref type="table" target="#tab_0">1</ref>, omitting any other information (for example, which tutor was responsible for a student). The results reported are based on re-classification (we reserve a cross-validation like mechanism for the more detailed experiments later on) and are shown in Table <ref type="table" target="#tab_3">4</ref>.</p><p>What does a difference signify? To answer that, one can take a step backwards and try to answer a simpler question: what does a large difference signify? We have elected to brand a difference as large when the re-classification accuracy of the same module for the same year differs by at least 20 percentage points when we compare the model predicting the pass/fail result of the first stage of the final exam and the corresponding model after a possible second stage (which is the actual pass/fail grade for the module). In Table <ref type="table" target="#tab_3">4</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>such differences are shown in bold.</head><p>There are two issues that become apparent when one views Table <ref type="table" target="#tab_3">4</ref>. The first is that whenever we observe an increase in the model accuracy when switching from the first exam (E) to the final grade (F), this is associated with senior modules where eventual success rates (see Table <ref type="table" target="#tab_1">2</ref>) are substantial. The only decrease is observed in a junior year module where success rates are considerably reduced compared to senior year modules. It is straightforward to attribute the increase in senior year modules to the fact that, eventually, students have to focus on their exam and pass the test, regardless of how well they did along the year. The large discrepancy, however, suggests that the exercises do not serve well their goal, which is to keep the students engaged in the learning process. One could say that exercises are less of learning opportunities and more of necessary evils.</p><p>The dramatic decrease in the 2006-7 year results of the INF20 module are quite interesting. They reflect, basically, a huge fail rate in the first stage of the exam, which is well served by a small model that predicts failure all around.</p><p>When seen from that viewpoint, however, the relatively narrow margins of the junior year modules seem quite impressive, since they are also associated with low overall pass rates. The difference, however, is that the junior modules also report significant dropout rates which skews pessimistically the rates reported in Table <ref type="table" target="#tab_1">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Detecting tutor influence</head><p>If we take the data sets that were used in section 4.1 and put back in the information on which tutor was responsible for each student group, we can run the same experiments and try to see whether the tutor attribute will surface in some models (sample data are shown in Table <ref type="table" target="#tab_4">5</ref>).</p><p>In principle, observing models where the tutor attribute appears near the decision tree root would not be a good thing, suggesting that a crucial factor in student success is not the educational system itself but the tutor. As a matter of fact we can opt to not look for this information at all in the resulting trees; comparing the accuracies to the ones reported in Table <ref type="table" target="#tab_3">4</ref> should suffice. These results are now shown in Table <ref type="table" target="#tab_5">6</ref>. Furthermore, we tried to summarise the results from a further point of view: that of consistency between the results reported for the E and F columns of both tables. Essentially we computed the quantity (F 5 -E 5 )-(F 6 -E 6 ) for each module for each year, where the subscript indicates which table that particular number was drawn from. Not surprisingly, the two singularities observed were module INF23 for year 2005-6 (with a value of about 20%) and module INF31 for year 2006-7 (with a value of about -22%).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Observing the accuracy-size trade-off</head><p>It is interesting to investigate whether the conventional wisdom on model characteristics is valid. In particular, we analysed the results in Table <ref type="table" target="#tab_5">6</ref> and in Table <ref type="table">7</ref> with respect to whether an increase (or decrease, accordingly) in model accuracy for a particular module for a year was associated with a reduction in model size. We say that the model accuracy increases if the accuracy for the E column of that year is less that the corresponding number in the F column. For the 68 pairs of numbers reported in Table <ref type="table" target="#tab_5">6</ref> and in Table <ref type="table">7</ref> we observed that only in 4 of them did we see the same direction in model accuracy and model size. So, conventional wisdom was confirmed in nearly 95% of the cases.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">DISCUSSION</head><p>HOU has been the first university in Greece to operate, from its very first year, a comprehensive assessment scheme (on tutoring and administrative services). Despite a rather hostile political environment (at least in Greece), quite a few academic departments have lately been moving along the direction of introducing such schemes, though the practice has yet to be adapted at a university level. Still, however, there is quite a mentality shift required when considering the subtle differences between "measuring" and "assessing".</p><p>The act of measuring introduces some error in what is being measured. If indices are interpreted as assessment indices, then people (actually, any "assessed" subject where people are involved -groups of people, for example) will gradually skew their behaviour towards achieving good measurements. Such behaviour is quite predictably human, of course; the problem is that it simply educates people in the ropes of the measurement system while sidelining the real issue of improving the educational service.</p><p>By shifting measurement to quantities that are difficult to "tweak", one hopes that people whose performance is assessed will gradually shift form fine-tuning their short-term behaviour toward achieving longer-term goals. Indeed, if people find out that the marginal gains from fine-tuning their behaviour are too small for the effort expended to achieve them, it will be easier to convince them to improve more fundamental attitudes towards tutoring (as far as tutors are concerned) or studying (as far as students are concerned).</p><p>In our application, this is demonstrated two-fold. First, by disseminating tutor group homogeneity indices, one hopes that, regardless how we call these indices, these tutor groups will be motivated by peer pressure to consider their performance vis-a-vis other tutor groups. Even if that may not be really required, that introspection itself will quite likely improve how hat particular tutor group co-operates; at least it will focus their decisions with respect to why such decisions might influence their overall ranking.</p><p>For students, a similar argument applies. Realising that one fits a model which predicts likely failure, even if one knows that the particular model is known to err quite some times, is something that will most likely motivate that person to take a more decisive approach to studying. For adult students, such a decisive approach might even mean to drop a course of studying or defer studying. This is not necessarily negative, however; knowing how to better utilise one's resources is a key skill in life long learning.</p><p>We have selected decision trees because we want to generate models that can be effectively communicated to tutors and students alike. We have also selected genetic algorithms to induce the decision trees because we have shown <ref type="bibr" target="#b6">[7]</ref> that, for the particular application domain, we can derive small and easy to communicate yet accurate trees. We thus need a hybrid approach: rule-based output to be comprehensible and grounded and evolutionary computing to derive this output.</p><p>Which other techniques should one utilise to develop the models? We cannot fail to note that conventional statistics can be cumbersome to disseminate to people with a background on humanities or arts, and this could have an adverse impact on the user acceptance of such systems. In that sense, the decision of whether the models are computed centrally or in a decentralized fashion (by devolving responsibility to the tutors, for example) is a key factor. In any case, deploying our measurement scheme in an organization-wide context would also lend support to our initial preference for short models. At the same time, the possibility of a decentralized scheme also suggests that we should strive to use tools that do not demand a steep learning curve on the part of the tutors.</p><p>Of course, one can take an alternative course and drop the requirement that a model has to be communicated. If we only focus on the indices then any technique can be used, from neural networks to more conventional ones, such as naive Bayes or logistic regression <ref type="bibr" target="#b3">[4]</ref>. As in all data mining application contexts, it is the application that must drive the techniques to use; for our problem, suffice to note that the comparisons reported (Table <ref type="table" target="#tab_3">4</ref>, Table <ref type="table" target="#tab_5">6</ref> and Table <ref type="table">7</ref>) are essentially technique independent, yet the GATREE approach has proven to-date to be the best method for prototyping the measurement exercise that we are developing.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">CONCLUSION</head><p>We have shown how we have used a combination of genetic algorithms and decision trees in the context of experimenting with how one might setup a quality control system in an educational context. Quality control should be a core aspect of any educational system but setting up a system for quality control entails managerial and administrative decisions that may also have to deal with political side-effects. Deciding how to best and as early as possible defuse the potential stand-offs that a quality measurement message might trigger calls for the employment of techniques that not only ensure a basic technical soundness in the actual measurement but also cater to the way the results are conveyed and subsequently exploited. This is particularly so when the application context for the large scale suggests that data and models will freely flow amongst thousands of tutors and tens of thousands of students.</p><p>We have earlier <ref type="bibr" target="#b8">[9]</ref> expressed the view that our approach is applicable to any educational setting where performance measurement can be cast in terms of test (and exam) performance. In the proposed paper we have scaled up our analysis to cover several modules and years and still believe that taking the sting out of individual performance evaluation but still being able to convey the full message is a key component of tutoring self-improvement. Scaling our approach to other programmes, other institutions and, even, obtaining the approval of our own university for official and consistent reporting of such indices is, however, less of a technical nature and more of a political exercise. After all we need to persuade people that some innovations are less of a threat and more of an opportunity.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 .</head><label>1</label><figDesc>Figure 1. A sample decision tree [3].</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>A sample decision tree training set (adapted from<ref type="bibr" target="#b2">[3]</ref>).</figDesc><table><row><cell>Assgn1</cell><cell>Assgn2</cell><cell>Assgn3</cell><cell>Assgn4</cell><cell>Exam</cell></row><row><cell>...</cell><cell>...</cell><cell>...</cell><cell>...</cell><cell>...</cell></row><row><cell>4.6</cell><cell>7.1</cell><cell>3.8</cell><cell>9.1</cell><cell>PASS</cell></row><row><cell>9.1</cell><cell>5.1</cell><cell>4.6</cell><cell>3.8</cell><cell>FAIL</cell></row><row><cell>7.6</cell><cell>7.1</cell><cell>5.8</cell><cell>6.1</cell><cell>PASS</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Success (percentage) rates of modules.</figDesc><table><row><cell></cell><cell>2004-5</cell><cell>2005-6</cell><cell>2006-7</cell></row><row><cell>INF10</cell><cell>35%</cell><cell>38%</cell><cell>33%</cell></row><row><cell>INF11</cell><cell>55%</cell><cell>52%</cell><cell>55%</cell></row><row><cell>INF12</cell><cell>39%</cell><cell>34%</cell><cell>35%</cell></row><row><cell>INF20</cell><cell>56%</cell><cell>44%</cell><cell>44%</cell></row><row><cell>INF21</cell><cell>37%</cell><cell>44%</cell><cell>37%</cell></row><row><cell>INF22</cell><cell>71%</cell><cell>61%</cell><cell>55%</cell></row><row><cell>INF23</cell><cell>N/A</cell><cell>83%</cell><cell>97%</cell></row><row><cell>INF24</cell><cell>70%</cell><cell>64%</cell><cell>58%</cell></row><row><cell>INF30</cell><cell>81%</cell><cell>85%</cell><cell>84%</cell></row><row><cell>INF31</cell><cell>93%</cell><cell>92%</cell><cell>85%</cell></row><row><cell>INF35</cell><cell>N/A</cell><cell>98%</cell><cell>93%</cell></row><row><cell>INF37</cell><cell>N/A</cell><cell>98%</cell><cell>98%</cell></row><row><cell>INF42</cell><cell>N/A</cell><cell>N/A</cell><cell>100%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 .</head><label>3</label><figDesc>Enrollment numbers at modules.</figDesc><table><row><cell>Assgn2 in [3..6]</cell><cell></cell></row><row><cell>FAIL</cell><cell>Assgn4 &lt; 3</cell></row><row><cell>FAIL</cell><cell>PASS</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>Model accuracies omitting tutor data.</figDesc><table><row><cell></cell><cell cols="2">2004-5</cell><cell cols="2">2005-6</cell><cell>2006-7</cell><cell></cell></row><row><cell></cell><cell>E</cell><cell>F</cell><cell>E</cell><cell>F</cell><cell>E</cell><cell>F</cell></row><row><cell>INF10</cell><cell>83</cell><cell>84</cell><cell>84</cell><cell>82</cell><cell>83</cell><cell>82</cell></row><row><cell>INF11</cell><cell>75</cell><cell>76</cell><cell>76</cell><cell>78</cell><cell>75</cell><cell>80</cell></row><row><cell>INF12</cell><cell>74</cell><cell>76</cell><cell>86</cell><cell>74</cell><cell>78</cell><cell>74</cell></row><row><cell>INF20</cell><cell>76</cell><cell>70</cell><cell>76</cell><cell>59</cell><cell>87</cell><cell>60</cell></row><row><cell>INF21</cell><cell>83</cell><cell>78</cell><cell>76</cell><cell>72</cell><cell>77</cell><cell>73</cell></row><row><cell>INF22</cell><cell>68</cell><cell>80</cell><cell>68</cell><cell>76</cell><cell>63</cell><cell>70</cell></row><row><cell>INF23</cell><cell>N/A</cell><cell></cell><cell>46</cell><cell>78</cell><cell>89</cell><cell>99</cell></row><row><cell>INF24</cell><cell>67</cell><cell>67</cell><cell>68</cell><cell>66</cell><cell>69</cell><cell>70</cell></row><row><cell>INF30</cell><cell>77</cell><cell>82</cell><cell>64</cell><cell>85</cell><cell>71</cell><cell>94</cell></row><row><cell>INF31</cell><cell>65</cell><cell>95</cell><cell>86</cell><cell>93</cell><cell>68</cell><cell>91</cell></row><row><cell>INF35</cell><cell>N/A</cell><cell></cell><cell>72</cell><cell>97</cell><cell>80</cell><cell>92</cell></row><row><cell>INF37</cell><cell>N/A</cell><cell></cell><cell>95</cell><cell>100</cell><cell>95</cell><cell>98</cell></row><row><cell>INF42</cell><cell>N/A</cell><cell></cell><cell>N/A</cell><cell></cell><cell cols="2">96 100</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 .</head><label>5</label><figDesc>An expanded sample training set (see Group).</figDesc><table><row><cell cols="2">Assgn1 Assgn2</cell><cell>Assgn3</cell><cell>Assgn4</cell><cell>Group</cell><cell>Exam</cell></row><row><cell>...</cell><cell>...</cell><cell>...</cell><cell>...</cell><cell>...</cell><cell>...</cell></row><row><cell>4.6</cell><cell>7.1</cell><cell>3.8</cell><cell>9.1</cell><cell>Athens-1</cell><cell>PASS</cell></row><row><cell>9.1</cell><cell>5.1</cell><cell>4.6</cell><cell>3.8</cell><cell>Patras-1</cell><cell>FAIL</cell></row><row><cell>7.6</cell><cell>7.1</cell><cell>5.8</cell><cell>6.1</cell><cell>Athens-2</cell><cell>PASS</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6 .</head><label>6</label><figDesc>Model accuracies including tutor data.</figDesc><table><row><cell></cell><cell cols="2">2004-5</cell><cell cols="2">2005-6</cell><cell cols="2">2006-7</cell></row><row><cell></cell><cell>E</cell><cell>F</cell><cell>E</cell><cell>F</cell><cell>E</cell><cell>F</cell></row><row><cell>INF10</cell><cell>82</cell><cell>83</cell><cell>80</cell><cell>79</cell><cell>82</cell><cell>81</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>INF37</cell><cell>N/A</cell><cell>95</cell><cell>98</cell><cell>91</cell><cell>95</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>INF42</cell><cell>N/A</cell><cell>N/A</cell><cell>93</cell><cell>100</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ACKNOWLEDGEMENTS</head><p>Thanassis Hadzilacos (now at the Open University of Cyprus) has contributed to this line of research while at Hellenic Open University.</p><p>Anonymized data can be available on request for research purposes only, on a case-by-case basis.</p><p>We acknowledge the advice, from an anonymous reviewer of the CIMA-ECAI08 workshop, on how to improve the presentation of this work to reflect the combination of AI techniques used.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This time we observe that the relative difference between the models which utilise the tutor attribute and the ones that do not are quite small. There are some very interesting cases, however.</p><p>For example, the INF11 module demonstrates near zero differences throughout. It is interesting to note that this module utilizes a plenary exam marking session, which means that tutors get to mark exam papers drawn from all groups at random. This places only marginal administrative overhead and, when viewed from the point of model consistency, seems to be well worth it.</p><p>Another example is the INF31 module (shown in bold), which demonstrated a year where the tutor attribute seemed to be of paramount importance. In that year, the gap between the first exam stage and the final grade seems to be influenced by the tutors. It is now very narrow (89 to 91) while it was quite wide (68 to 91). This could suggest a relative gap in tutor homogeneity.</p><p>There is one other way to view the importance of the tutor attribute. One can derive a model for one module group and then attempt to use that model as a predictor of performance for the other module groups (within the same module). This approach, while suppressing the tutor attribute, essentially tests its importance by specifically segmenting the module data set along groups. The overall accuracy is then averaged over all individual tests. This is the lesion comparison; its results are shown in Table <ref type="table">7</ref>.</p><p>We highlight (in bold) the main difference from the results in Table <ref type="table">4</ref>, where it now seems that the gap has been shortened a while. Surprisingly, it suggests an erratic intra-group consistency. Note also, that this particular result in Table <ref type="table">4</ref> was the only one not to pass the binary choice (50%) level, which it only just did in Table <ref type="table">6</ref>. </p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Machine Learning</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mitchell</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1997">1997</date>
			<publisher>McGraw Hill</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">C4.5: Programs for machine learning</title>
		<author>
			<persName><forename type="first">J</forename><surname>Quinlan</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1993">1993</date>
			<publisher>Morgan Kaufmann</publisher>
			<pubPlace>San Mateo, CA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Analyzing student performance in distance learning with genetic algorithms and decision trees</title>
		<author>
			<persName><forename type="first">D</forename><surname>Kalles</surname></persName>
		</author>
		<author>
			<persName><surname>Ch</surname></persName>
		</author>
		<author>
			<persName><surname>Pierrakeas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="655" to="674" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Predicting students&apos; performance in distance learning using Machine Learning techniques</title>
		<author>
			<persName><forename type="first">S</forename><surname>Kotsiantis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Pierrakeas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pintelas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="411" to="426" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Data mining: practical machine learning tools and techniques with Java implementations</title>
		<author>
			<persName><forename type="first">I</forename><surname>Witten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Frank</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2000">2000</date>
			<publisher>Morgan Kaufmann</publisher>
			<pubPlace>San Mateo, CA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Breeding decision trees using evolutionary techniques</title>
		<author>
			<persName><forename type="first">A</forename><surname>Papagelis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kalles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Machine Learning</title>
				<meeting>the International Conference on Machine Learning<address><addrLine>Williamstown, Massachusetts</addrLine></address></meeting>
		<imprint>
			<publisher>Morgan Kaufmann</publisher>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="393" to="400" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Using Genetic Algorithms and Decision Trees for a posteriori Analysis and Evaluation of Tutoring Practices based on Student Failure Models</title>
		<author>
			<persName><forename type="first">D</forename><surname>Kalles</surname></persName>
		</author>
		<author>
			<persName><surname>Ch</surname></persName>
		</author>
		<author>
			<persName><surname>Pierrakeas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 3 rd IFIP conference on Artificial Intelligence Applications and Innovations</title>
				<meeting>the 3 rd IFIP conference on Artificial Intelligence Applications and Innovations<address><addrLine>Athens, Greece</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="9" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">On Small Data Sets Revealing Big Differences</title>
		<author>
			<persName><forename type="first">Th</forename><surname>Hadzilacos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kalles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ch</forename><surname>Pierrakeas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Xenos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 4th Panhellenic conference on Artificial Intelligence</title>
		<title level="s">Springer LNCS</title>
		<meeting>the 4th Panhellenic conference on Artificial Intelligence<address><addrLine>Heraklion, Greece</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
			<biblScope unit="volume">3955</biblScope>
			<biblScope unit="page" from="512" to="515" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">On the Software Engineering Aspects of Educational Intelligence</title>
		<author>
			<persName><forename type="first">Th</forename><surname>Hadzilacos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kalles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th International Conference on Knowledge-Based Intelligent Information &amp; Engineering Systems</title>
		<title level="s">Springer LNCS</title>
		<meeting>the 10th International Conference on Knowledge-Based Intelligent Information &amp; Engineering Systems<address><addrLine>Bournemouth, UK</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
			<biblScope unit="volume">4252</biblScope>
			<biblScope unit="page" from="1136" to="1143" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">A survey on student dropout rates and dropout causes concerning the students in the Course of Informatics of the Hellenic Open University</title>
		<author>
			<persName><forename type="first">M</forename><surname>Xenos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Pierrakeas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pintelas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computers &amp; Education</title>
		<imprint>
			<biblScope unit="volume">39</biblScope>
			<biblScope unit="page" from="361" to="377" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
