1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
|
\section{Learning rules}
\label{sec:rules}
The goal of learning rules in this paper is to discover and explain common approaches and mistakes in student programs. We use a rule learner called ABCN2e, implemented within the Orange data mining library~\cite{demsar2013orange}. ABCN2e modifies\footnote{Modifications are described in a technical report available at \url{https://ailab.si/abml}.} the original CN2 algorithm~\cite{clarkECML1991}.
General rule-learning algorithms, such as CN2, tend to generate many specific rules. This produces more accurate results but makes rules harder to explain. This section describes the problem-specific configuration of the rule-learning algorithm for extracting relevant and explainable patterns from student programs.
Each program is represented in the feature space of AST patterns described in the previous section. Based on test results each program is classified either as \emph{correct} or \emph{incorrect}. Our main motivation for learning rules bases on assumption that the patterns highly correlated with incorrect programs represent mistakes.
A program can be incorrect for one of two reasons: either a) it contains some incorrect pattern (a buggy pattern) that should be removed or modified, or b) it is missing one or more programing constructs (patterns) that should be present for the program to be correct.
Classification rules can express both reasons. To identify buggy patterns we learn a set of rules covering incorrect programs, where each condition in the rule must express the presence of a pattern. For missing patterns, we learn another set of rules, which may contain missing patterns within their conditions, and describe the missing constructs in a program that have to be implemented. All rules explaining incorrect programs are called \emph{n-rules}.
To learn explainable, meaningful and non-redundant rules, we impose the following additional constraints on the rule learner:
\begin{itemize}
\item classification accuracy of each rule must exceed 90\%, because we deem a 10\% false-positive error as acceptable;
\item each term in the condition of a rule must be significant according to the likelihood test, meaning that each pattern in the condition part is indeed relevant (we set the significance threshold to p=0.05);
\item a condition can have at most 3 patterns; and
\item each rule must cover at least 5 distinct programs -- to avoid learning redundant rules representing the same error.
\end{itemize}
To identify different approaches to solving an exercise, we learn rules that explain correct programs. A program is correct when it implements all required patterns and no buggy patterns. There may be several possible sets of required patterns for each exercise, with each set corresponding to a different approach.
We use the same constraints as in the case of n-rules and learn rules for correct programs called \emph{p-rules}. In this case, we always require that conditions mention the presence of patterns, since it is easier to explain different approaches of students with something they have written and not with something they have not. To account for possible buggy patterns, the requirement to achieve 90\% classification accuracy was not evaluated on full data, but only on data not covered by n-rules. Hence, a rule can cover an example with a specific approach even though it contains a buggy pattern.
\section{Interpreting rules}
\label{sec:interpreting-rules}
Learned rules can be used to analyze student programing. This section describes several rules induced for two Python exercises: \textsf{Fahrenheit to Celsius}, which reads a value from standard input and calculates the result, and \textsf{Greatest Absolutist}, one of the introductory exercises for functions.
\subsection{Fahrenheit to Celsius}
The first problem in CodeQ Python course is to write a program converting from degrees Fahrenheit to degrees Celsius. The program should ask the user to input a temperature, and print the result. A sample correct program is:
\begin{Verbatim}
F = float(input("Fahrenheit: "))
C = 5 / 9 * (F - 32)
print("Celsius: ", C)
\end{Verbatim}
Students have submitted 1177 programs for this problem, with 495 correct and 682 incorrect programs. Our systems extracted 891 relevant AST patterns, which were used as attributes in rule learning. The rule learner induced 24 n-rules, 14 of which mention only presence of patterns, and 16 p-rules.
We first take a look at n-rules that mention only presence of patterns in their conditions. The most accurate rule according to the rule learner was:
\begin{Verbatim}[fontfamily=sf]
P20 ⇒ incorrect [208, 1]
\end{Verbatim}
\noindent
This rule covers programs where the pattern \textsf{P20} is present. It implies an incorrect program, and covers 208 incorrect and one correct program. \textsf{P20} is the AST pattern describing a call to the \texttt{int} function:
\begin{Verbatim}[fontfamily=sf]
(Module (body (Assign (value (Call (func (Name (id int) (ctx Load))))))))
\end{Verbatim}
The second best n-rule covers 72 incorrect and no correct programs:
\begin{Verbatim}[fontfamily=sf]
P5 ∧ P35 ⇒ incorrect [72, 0]
\end{Verbatim}
\noindent
Pattern \textsf{P5} matches programs where the result of the \texttt{input} call is not cast to \texttt{float} but stored as a string. Pattern \textsf{P35} matches programs where the value 32 is subtracted from a variable on the left-hand side of a multiplication. Sample programs matching the first rule (left) and the second rule (right) are:
\begin{Verbatim}
g2 = input() g2 = \blue{\underline{input}}('Temperature [F]? ')
g1 = \blue{\underline{int}}(g2) g1 = (\red{\dashuline{(g2 - 32) *}} (5 / 9))
print(((g1-32)*(5/9))) print(g2, 'F equals', g1, 'C')
\end{Verbatim}
These rules describe two common student errors. The left program is incorrect, since it fails when the user inputs a decimal. The right program is incorrect because the input string must be cast to a number. Not casting it (pattern \textsf{P5}) and then using it in an expression (pattern \textsf{P35}) will raise an exception.
The two most accurate n-rules with missing patterns in their conditions are:
\begin{Verbatim}[fontfamily=sf]
¬P0 ⇒ incorrect [106, 0]
¬P1 ∧ P16 ⇒ incorrect [100, 0]
\end{Verbatim}
\noindent
Pattern \textsf{P0} matches programs with a call to function \texttt{print}. A program without a \texttt{print} is always incorrect, since it will not output anything.
The second rule covers programs with \textsf{P1} missing and \textsf{P16} present. \textsf{P16} matches programs with a call to the \texttt{print} function, where the argument contains a formula which subtracts 32 from a variable and then further multiplies the result. \textsf{P1} describes a call to the function \texttt{float} as the first item in an expression, i.e. \texttt{= float(...)}. This rule therefore represents programs that have the formula in the \texttt{print} function (\textsf{P16} is present), however fail to cast input from string to float (\textsf{P1} is missing).
Let us now examine the other type of rules. The best four p-rules are:
\begin{Verbatim}[fontfamily=sf]
P2 ∧ P8 ⇒ correct [1, 200]
P1 ∧ P42 ⇒ correct [0, 68]
P1 ∧ P8 ⇒ correct [3, 217]
P80 ⇒ correct [0, 38]
\end{Verbatim}
\noindent
Patterns in the condition of the first rule, \textsf{P2} and \textsf{P8}, correspond respectively to expressions of the form \texttt{float(input(?))} and \texttt{print((?-32)*?)}. Programs matching both patterns wrap the function \texttt{float} around \texttt{input}, and have an expression that subtracts 32 and then uses multiplication within the \texttt{print}.
This first rule demonstrates an important property of p-rules: although patterns \textsf{P2} and \textsf{P8} are in general not sufficient for a correct program (it is trivial to implement a matching but incorrect program), only one out of 201 student submissions matching these patterns was incorrect. This suggests that the conditions of p-rules represent the critical elements of the solution. Once students have figured out these patterns, they are almost certain to have a correct solution. A sample program matching the first rule is:
\begin{Verbatim}
g1 = \blue{\underline{float(input(}}'Temperature [F]: '))
print(((g1 \red{\dashuline{- 32) *}} (5 / 9)))
\end{Verbatim}
\noindent
The second and third p-rules are variations of the first. For instance, the second rule describes programs that have the formula in the argument to the \texttt{print} function. The fourth rule, however, is different. \textsf{P80} describes programs that subtract 32 from a variable cast to float. The following program matches \textsf{P80}:
\begin{Verbatim}
g1 = input('Fahrenheit?')
g0 = (\blue{\underline{(float(g1) - 32)}} * (5 / 9))
print(g0)
\end{Verbatim}
\subsection{Greatest Absolutist}
In this exercise students must implement a function that accepts a list of numbers and returns the element with the largest absolute value. One solution is
\begin{Verbatim}[fontfamily=tt]
def max_abs(l):
vmax = l[0]
for v in l:
if (abs(v) > abs(vmax)):
vmax = v
return vmax
\end{Verbatim}
We have received 155 submissions (57 correct, 98 incorrect) for this exercise. Due to its higher complexity and since the solutions contain definitions of lists, we obtained 8298 patterns to be used as attributes in learning. High number of patterns together with a low number of learning examples could present a problem for rule learning, since the large number of candidate rules could lead to statistical anomalies. Still, most of the learned rules seem valuable.
The rule-learning algorithm learned 15 n-rules (7 mentioning only presence of patterns) and 6 p-rules. Below we can see the two best n-rules referring to the presence of patterns and two programs (left for first rule, right for second):
\begin{Verbatim}[fontfamily=sf]
P64 ⇒ incorrect [22, 0]
P2 ∧ P70 ⇒ incorrect [17, 0]
\end{Verbatim}
\begin{Verbatim}
def max_abs(l): def max_abs(l):
vmax = 0 vmax = None
for i in range(len(l)): for v in l:
if \blue{\underline{vmax}} < abs(l[i]): if vmax==None or vmax<v:
vmax = l[i] \red{\dashuline{vmax}} = \blue{\underline{abs}}(v)
return \blue{\underline{vmax}} return \red{\dashuline{vmax}}
\end{Verbatim}
The pattern from the first rule, \textsf{P64}, matches functions returning the variable that is used in the condition of an if clause without an application of another function (such as \texttt{abs}). The left program contains this pattern, where the value \texttt{vmax} is compared in the if clause and then returned. This error is quite common among students, because they forget to compare the absolute value of \texttt{vmax}.
The second rule contains two patterns. \textsf{P70} (blue underlined) matches functions containing the call to \texttt{abs} in an assignment statement nested within a for loop and an if clause. \textsf{P2} (red dash) matches functions that return the variable used in an assignment statement within a for-if block. Such programs are incorrect since they do not store the original list element. For example, if -7 has the largest absolute value in the list, then the function should return -7 and not 7.
The best two n-rules with absence of patterns in condition are:
\begin{Verbatim}[fontfamily=sf]
P1 ∧ ¬P11 ∧ ¬P131 ⇒ incorrect [34, 0]
P36 ∧ ¬P162 ⇒ incorrect [26, 0]
\end{Verbatim}
\noindent
The first rule covers programs matching \textsf{P1} (checks for a function definition in the program) but missing \textsf{P11} (if the iteration variable in a for loop is directly assigned to another variable within an if clause) and \textsf{P131} (whether the return statement uses indexing, i.e. \texttt{return l[?]}). One such example is the above right program: it contains a function definition, it does not directly assign the value of \texttt{v} but uses its absolute value, and does not use indexing in the return statement.
This rule specifies two missing patterns, which makes is harder to understand. If one of the two missing patterns were implemented, the rule would not cover this program any more. Therefore, the questions is, which of these two reasons is really missing? Different missing patterns could be understood as different options to finalize the program.
The second rule identifies one missing pattern. \textsf{P36} matches a call to function \texttt{max} in the return statement, whereas \textsf{P162} matches a call to \texttt{max} with the given list as the argument. This rule covers, for example, the following program:
\begin{Verbatim}
def max_abs(l):
return max(abs(l))
\end{Verbatim}
\noindent
It uses \texttt{max} in the return statement, but does not apply it directly to the list \texttt{l}.
The four most accurate p-rules induced by our rule learner were:
\begin{Verbatim}[fontfamily=sf]
P11 ∧ P17 ∧ P35 ⇒ correct [0, 20]
P11 ∧ P27 ∧ P3 ⇒ correct [2, 34]
P519 ⇒ correct [0 9]
P27 ⇒ correct [ 6 38]
\end{Verbatim}
\noindent
Sample programs covered by the first and second rules are:
\begin{Verbatim}[fontfamily=tt]
def max_abs(l): def max_abs(\green{\dotuline{l}}):
\green{\dotuline{vmax = 0}} vmax = l[0]
for \blue{\underline{v}} in l: for \blue{\underline{v}} in \green{\dotuline{l}}:
if abs(\red{\dashuline{vmax}}) < abs(v): if abs(\red{\dashuline{vmax}}) < abs(v):
\red{\dashuline{vmax}} = \blue{\underline{v}} vmax = \blue{\underline{v}}
return vmax return \red{\dashuline{vmax}}
\end{Verbatim}
\noindent
The first two rules and the above programs are similar. Both rules share a common reason, \textsf{P11} (blue underlined in both programs), describing a pattern, where the variable from the for loop is used in the right side of an assignment within the if clause. \textsf{P17} and \textsf{P27} are also similar (red dash underline). The former links the occurrence of a variable within the \texttt{abs} function in an if condition with the variable from an assignment, whereas the latter links the same variable from an if condition with the variable from the return statement. \textsf{P35} matches variable assignments to 0, hence the first rule covers solutions initializing \texttt{vmax} to zero. \textsf{P3} matches for-looping over the input list.
After inspecting all covered examples of the first and the second rule, we discovered that the first rule is only a more strict version of the second rule, since all examples covered by the first rule are also covered by the second rule. These two rules therefore do not describe two different approaches, but a more general and a more specific representation of the same approach. Similarly, the fourth rule is a generalization of the first two rules, containing only \textsf{P27} within conditions. It appears that the pattern \textsf{P27} is the most important; of 44 programs, where students used the absolute value of \texttt{vmax} in comparison and returned \texttt{vmax} at the end, 38 were evaluated as correct.
The third rule describes a different pattern. It covers programs that define a list containing values 2, 1, and -6. Defining such a list is evidently not necessary for the solution of this exercise. Why would it then correlate with the correctness of the solution?
To explain this rule we first have to describe how students test their programs. One option is to simply use the \textit{Test} button, which submits the program to a server, where it is tested against a predefined set of test cases. The other option is to click the \textit{Run} button, which runs the program and outputs the results. Those students who defined a list with values 2, 1, and -6 in their programs are most likely using the second option. They create their own test cases and then submit a program only when they are certain that it is correct. Since the description of the exercise includes a single test case with values 2, 1, and -6, most students use this list as the testing case. On the other hand, given that the rule covers only nine programs, the probability that the rule is a statistical artifact is not negligible.
|