This page shows all 55 questions in the research agenda.
You can also see the questions arranged into Themes
Code | Question | Contributors |
---|---|---|
Q1 | What common errors do students make when answering online assessment questions? | |
Q2 | Do the errors students make in e-assessments differ from those they make in paper-based assessments? | |
Q3 | What are the approaches to detecting and feeding back on students’ errors? | |
Q4 | How can content-specific features of provided feedback, for instance explanations with examples versus generic explanations, support students' learning? | |
Q5 | What are the linguistic features of feedback that help students engage with and use feedback in an online mathematical task at hand and in future mathematical activities? | |
Q6 | What difficulties appear when designing e-assessment tasks that give constructive feedback to students? | |
Q7 | How can feedback that is dynamically tailored to the student’s level of mathematical expertise help a student use feedback on mathematical tasks effectively? | |
Q8 | How useful for students’ long-term learning is feedback that gives a series of follow-up questions, from a decision tree, versus a single terminal piece of feedback? | |
Q9 | What are the relative benefits of e-assessment giving feedback on a student’s set of responses (e.g. “two of these answers are wrong – find which ones and correct them”), rather than individual responses separately? | |
Q10 | Under what circumstances is diagnosing errors worth the extra effort, as compared with generally addressing errors known to be typical? | |
Q11 | What are the relative merits of addressing student errors up-front in the teaching compared with using e-assessment to detect and give feedback on errors after they are made? | |
Q12 | In what circumstances is instant feedback from automated marking preferable to marking by hand? | |
Q13 | How do students interact with an e-assessment system? | |
Q14 | To what extent does repeated practice on randomized e-assessment tasks encourage mathematics students to discover deep links between ideas? | |
Q15 | How do students engage with automated feedback? What differences (if any) can be identified with how they would respond to feedback from a teacher? | |
Q16 | What should students be encouraged to do following success in e-assessment? | |
Q17 | What are students' views on e-assessment, and what are their expectations from automated feedback? | |
Q18 | How might dyslexic, dyscalculic and other groups of students be disadvantaged by online assessments rather than paper-based assessments? | |
Q19 | How can peer assessment be used as part of e-assessment? | |
Q20 | How can e-assessment be used in group work, and what effect does the group element have on individuals' learning? | |
Q21 | What design methodologies and principles are used by e-assessment task designers? | |
Q22 | What principles should inform the design of e-assessment tasks? | |
Q23 | E-assessment task designers often convert questions that could be asked on a traditional pen and paper exam: what are the implications, technicalities, affordances and drawbacks of this approach? | |
Q24 | To what extent does the randomisation of question parameters, which makes sharing answers between students difficult, adequately address plagiarism? | |
Q25 | What effect does the use of random versions of a question (e.g. using parameterised values) have on the outcomes of e-assessment? | |
Q26 | When writing multiple choice questions, is student learning better enhanced using distractors based on common errors, or randomly-generated distractors? | |
Q27 | How can formative e-assessments improve students’ performance in later assessments? | |
Q28 | How can regular summative e-assessments support learning? | |
Q29 | What are suitable roles for e-assessment in formative and summative assessment? | |
Q30 | To what extent does the timing and frequency of e-assessments during a course affect student learning? | |
Q31 | What are the relations between the mode of course instruction and students' performance and activity in e-assessment? | |
Q32 | What advice and guidance (both practical and pedagogical) is available to lecturers about using e-assessment in their courses, and to what extent do they engage with it? | |
Q33 | What might a “hierarchy of needs” look like for lecturers who are transitioning to increased use of e-assessments? | |
Q34 | How can lecturers be informed about how students interact with e-assessment tasks, and so help lecturers act upon these findings in an effective way? | |
Q35 | What types of reasoning are required to complete current e-assessments? | |
Q36 | To what extent do existing e-assessments provide reliable measures of mathematical understanding, as might otherwise be measured by traditional exams? | |
Q37 | How can e-assessment support take-home open-book examinations? | |
Q38 | What developments at the forefront of e-assessment (such as artificial intelligence) can we apply to undergraduate mathematics? | |
Q39 | What methods are available for student input of mathematics? | |
Q40 | How can the suitability of e-assessment tools for summative assessment be improved by combining computer-marking and pen-marking? | |
Q41 | Are there differences in performance on mathematics problems presented on paper versus as e-assessments? | |
Q42 | How can we automate the assessment of work traditionally done using paper and pen? | |
Q43 | How can we emulate human marking of students’ working such as follow-on marking and partially correct marking? | |
Q44 | How can e-assessment using comparative judgment support learning? | |
Q45 | How can comparative judgement be used for e-assessment? | |
Q46 | How can we assess problem solving using e-assessment? | |
Q47 | How can we assess open-ended tasks using e-assessment? | |
Q48 | How can e-assessments provide scaffolding (cues, hints) during and after problem-solving tasks? | |
Q49 | How can the assessment of proof be automated? | |
Q50 | What can automated theorem provers (e.g. LEAN) offer to the e-assessment of proof comprehension? | |
Q51 | What types/forms of proof-comprehension-related questions can be meaningfully assessed using currently available e-assessment platforms? | |
Q52 | How can students effectively type free-form proof for human marking online? | |
Q53 | How can e-assessments be designed to expand and enrich students' example spaces? | |
Q54 | To what extent can e-assessments meaningfully judge student responses to example generation tasks? | |
Q55 | How does the use of e-assessment impact students’ example generation strategies and success, relative to the same tasks on paper or orally? |