summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorTimotej Lazar <timotej.lazar@fri.uni-lj.si>2017-01-18 17:52:53 +0100
committerTimotej Lazar <timotej.lazar@fri.uni-lj.si>2017-01-18 18:34:05 +0100
commit101fcfc37a5f99cf5add7fd2f9e13c2e65c4b27a (patch)
treea89ec88c4c08544361975555d45c5c1388f31220
parentd0f454b7346b2b7b5cfe13e97a32924bc94046bd (diff)
Add outline and files for the AIED2017 paper
-rw-r--r--aied2017/aied2017.tex23
-rw-r--r--aied2017/background.tex1
-rw-r--r--aied2017/conclusion.tex1
-rw-r--r--aied2017/dataset.tex1
-rw-r--r--aied2017/evaluation.tex1
-rw-r--r--aied2017/introduction.tex1
-rw-r--r--aied2017/method.tex1
-rw-r--r--aied2017/outline.org81
8 files changed, 110 insertions, 0 deletions
diff --git a/aied2017/aied2017.tex b/aied2017/aied2017.tex
new file mode 100644
index 0000000..d342a9c
--- /dev/null
+++ b/aied2017/aied2017.tex
@@ -0,0 +1,23 @@
+\documentclass{llncs}
+
+\usepackage[utf8]{inputenc}
+
+\begin{document}
+
+\title{TODO}
+\author{TODO}
+\institute{University of Ljubljana, Faculty of Computer and Information Science, Slovenia}
+\maketitle
+
+\begin{abstract}
+TODO
+\end{abstract}
+
+\input{introduction}
+\input{background}
+\input{dataset}
+\input{method}
+\input{evaluation}
+\input{conclusion}
+
+\end{document}
diff --git a/aied2017/background.tex b/aied2017/background.tex
new file mode 100644
index 0000000..563e5ea
--- /dev/null
+++ b/aied2017/background.tex
@@ -0,0 +1 @@
+\section{Background} \ No newline at end of file
diff --git a/aied2017/conclusion.tex b/aied2017/conclusion.tex
new file mode 100644
index 0000000..47c04f1
--- /dev/null
+++ b/aied2017/conclusion.tex
@@ -0,0 +1 @@
+\section{Conclusion} \ No newline at end of file
diff --git a/aied2017/dataset.tex b/aied2017/dataset.tex
new file mode 100644
index 0000000..e8d168d
--- /dev/null
+++ b/aied2017/dataset.tex
@@ -0,0 +1 @@
+\section{Dataset} \ No newline at end of file
diff --git a/aied2017/evaluation.tex b/aied2017/evaluation.tex
new file mode 100644
index 0000000..22e4253
--- /dev/null
+++ b/aied2017/evaluation.tex
@@ -0,0 +1 @@
+\section{Evaluation} \ No newline at end of file
diff --git a/aied2017/introduction.tex b/aied2017/introduction.tex
new file mode 100644
index 0000000..ac26833
--- /dev/null
+++ b/aied2017/introduction.tex
@@ -0,0 +1 @@
+\section{Introduction} \ No newline at end of file
diff --git a/aied2017/method.tex b/aied2017/method.tex
new file mode 100644
index 0000000..3a14d53
--- /dev/null
+++ b/aied2017/method.tex
@@ -0,0 +1 @@
+\section{Method} \ No newline at end of file
diff --git a/aied2017/outline.org b/aied2017/outline.org
new file mode 100644
index 0000000..a70c1a3
--- /dev/null
+++ b/aied2017/outline.org
@@ -0,0 +1,81 @@
+* Introduction (1) :Timotej:
+** Generics
+ITSs are useful, but expensive to create. Internet increases the reach of ITSs (think MOOC) and allows us to gather a lot of data. The challenge is to extract useful information from this data to improve feedback to students.
+ - koedinger2013new
+
+** Programming ITS
+A particularly challenging domain:
+ - no meaningful “solutions steps”
+ - no easily discernible and independent “program features”
+
+** Problem
+Find good features to describe student programs that:
+ - are automatically learnable
+ - can serve as a basis for generating hints
+
+** Contributions
+A pattern language for describing program features.
+Rules for classifying correct/incorrect Prolog programs.
+
+* Background (2) :Timotej:
+Model-tracing is not well-suited for programming tutors.
+ - Hint Factory (Barnes & Stamper) for programming (Jin, Rivers)
+ - student traces are basically insane anyway, so learning from them is not a good idea (Piech 2015)
+Instead we focus on analyzing individual submissions - CBM (maybe just /like/ CBM?).
+ - why: human tutors are usually able to comment without referring to previous program versions
+
+Analyzing program based on formal descriptions (usually defined manually):
+ - plans (Johnson)
+ - techniques (Brna/Huang/Le)
+ - strategies (Gerdes)
+Comparing programs directly / after canonicalization (Rivers 2015)
+ - not easy to determine which differences are crucial
+
+Defining “programming attributes/features”:
+ - AST subtrees (Nguyen 2014)
+ - linkage graphs (Jin 2012)
+ - n-grams or other purely textual attributes
+ - mostly used for plagiarism (MOSS) / code clone detection
+
+Tgrep / trx for structural patterns in NLP - use the same for describing code!
+
+* Dataset (3) :Timotej:
+Solution traces
+** Motivating example: ancestor/2
+ - show and explain correct / buggy program
+ - examples of patterns & rules
+ - buggy rules ≈ error classes
+ - correct rules ≈ program strategies or “plans” → determine intent
+ - how correct / buggy rules can be used to automatically determine intent and generate feedback
+ - using rules to assist authoring tutors
+
+** Patterns
+Programming is all about manipulating data (variables/literals).
+Patterns encode (only) relations between data nodes in an AST, and are more granular than subtrees.
+Patterns abstract away some information
+Algorithm for extracting patterns.
+ - which patterns are considered
+
+* Method (2) :Martin:
+Rule learning.
+ - first learn correct, then buggy rules
+Automatic hint generation.
+
+* Evaluation (3) :Martin:Timotej:
+Example: describe patterns/rules for ancestor/2 (and/or another domain).
+ - example trace
+CA/AUC for rules/RF/….
+ Using e.g. RF with patterns for attributes achieves a high classification accuracy for correct/incorrect programs.
+ Rules are less accurate, but can be used for hints (how can we improve this)?
+Simulated hints: for how many submissions would we suggest a hint that was ultimately implemented in the final solution?
+ Hints from buggy rules are almost always implemented; correct rules are less certain.
+
+* Conclusions / future work (1)
+ - syntactic variations (e.g. using = instead of implicit matching)
+ - combine with canonicalization
+ - ABML for rule learning
+ - integrate into a tutor (CodeQ / CloudCoder) to test in situ
+
+* Misc
+Problems:
+ - predicates with similar clauses (e.g. union/3)