Marking algorithms#
Every question part has a marking algorithm, which is is responsible for:
Rejecting the student’s answer if it’s invalid. If the answer is rejected, no credit or feedback will be given and the student must change their answer before resubmitting.
If the student’s answer is valid, assigning credit and giving feedback messages.
The credit for a part is the proportion of the marks available which should be awarded to the student. The total marks available are set by the question author, and might be reduced if the student reveals steps, or if this part is a gap in a gap-fill part.
The feedback messages shown to the student are strings of text shown after the part has been marked. These might only become visible to the student after they have finished the exam, so don’t rely on feedback messages to convey information that students might need in subsequent parts.
The marking algorithm comprises a set of marking notes written in JME syntax, which are evaluated similarly to question variables.
Two marking notes are required:
- mark#
The
mark
note should award credit and provide feedback based on the student’s answer. If the student’s answer is invalid,mark
shouldfail
.
- interpreted_answer#
The
interpreted_answer
note should produce a value representing the student’s answer to this part, which can be used by other parts with adaptive marking.
Each note evaluates to a value, and also produces a list of feedback items, which modify the amount of credit awarded or give a message to the student. When a feedback item modifies the amount of credit awarded, a message describing the number of marks awarded or taken away from the previous total is displayed to the student.
If a note fails, either because it applies the fail
function or an error is thrown while it is evaluated, it will produce no value and no feedback items.
Any notes referring to a failed note also fail.
If the mark
or interpreted_answer
notes fail, the student’s answer is rejected and the student must change their answer before resubmitting.
Like question variables, marking notes can refer to each other.
When another note is referred to in another note’s definition, its value is substituted in.
To apply another note’s feedback items, use apply
.
Variables available to marking algorithms#
The following variables are available for use by marking algorithms:
- path#
The path to this part, in the form
pN(gN|sN)
. The first part (part a) has pathp0
. As an example, the second gap in part c would have pathp2g1
.
- studentAnswer#
The student’s answer to the part. The data type of this value depends on the part of the type. See the list of standard part type values, and custom part type answer input methods for details on the data types produced by different part types.
- settings#
A
dict
of the part’s settings.For built-in parts, see the relevant part type’s documentation. For custom part types this is all of the settings defined in the part.
- gaps#
A
list
of the gaps belonging to this part. Each element in the list is adict
of the same variables that would be available in the gap’s own marking algorithm.
- steps#
A
list
of the steps belonging to this part. Each element in the list is adict
of the same variables that would be available in the gap’s own marking algorithm.
- input_options#
(Only for custom part types)
A
dict
of the options for the part’s answer input method.
Marking-specific JME functions#
All the built-in JME functions are available in marking notes, as well as the following functions specifically to do with marking:
- correct(message)#
Set the credit to 1 and give the feedback message
message
. Ifmessage
is omitted, the default “Your answer is correct” message for the current locale is used.
- incorrect(message)#
Set the credit to 0 and give the feedback message
message
. Ifmessage
is omitted, the default “Your answer is incorrect” message for the current locale is used.
- correctif(condition)#
If
condition
evaluates totrue
, set the credit to 1 and give the default feedback message. Otherwise, set the credit to 0 and give the default feedback message.Equivalent to
if(condition,correct(),incorrect())
.
- set_credit(credit, message)#
Set the credit to
credit
, and give the feedback messagemessage
. The message should explain why the credit was awarded.
- add_credit(credit, message)#
Add
credit
to the current total, to a maximum of 1, and give the feedback messagemessage
. The message should explain why the credit was awarded.If
credit
is negative, credit is taken away, to a minimum of 0.
- add_credit_if(condition, credit, positive_message, negative_message)#
If
condition
evaluates totrue
, addcredit
to the current total, and give the feedback messagepositive_message
.If
condition
evaluates to false, show the feedback messagenegative_message
. Ifcredit > 0
, then this is considered to be negative feedback, otherwise it is neutral.You can omit
negative_feedback
, in which case no feedback will be shown whencondition
is false.
- sub_credit(credit, message)#
Subtract
credit
from the current total and give the feedback messagemessage
. The message should explain why the credit was taken away.
- multiply_credit(proportion, message)#
Multiply the current credit by
proportion
and give the feedback messagemessage
. The message should explain why the credit was modified.This operation is displayed to the student as an absolute change in marks awarded, not a multiplication. For example, if the student already had 2 marks and multiply_credit(0.5,message) was applied, the message displayed would be along the lines of “1 mark was taken away”.
- multiply_credit_if(condition, proportion, positive_message, negative_message)#
If
condition
evaluates totrue
, multiply the current credit by``proportion``, and give the feedback messagepositive_message
.If
condition
evaluates to false, show the feedback messagenegative_message
.You can omit
negative_feedback
, in which case no feedback will be shown whencondition
is false.
- end()#
End the marking here. Any feedback items produced after this one are not applied.
This is most useful as a way of stopping marking once you’ve decided the student’s answer is incorrect partway through a multi-step marking process.
- fail(message)#
Reject the student’s answer as invalid, set the credit to 0 and give the feedback message
message
. The message should explain why the student’s answer was rejected.The marking ends here.
Since the student might not see the feedback message until the exam is over, you should also use
warn
to add a warning message next to the input field describing why the student’s answer was rejected.
- warn(message)#
Show a warning next to the answer input. This does not affect credit or stop the running of the marking algorithm.
- feedback(message)#
Give the feedback message
message
, without modifying the credit awarded.
- positive_feedback(message)#
Give the feedback message
message
, without modifying the credit awarded, but with a positive annotation (a green tick in the default theme).
- negative_feedback(message)#
Give the feedback message
message
, without modifying the credit awarded, but with a negative annotatin (a red cross in the default theme).
- x ; y#
Add feedback items generated by
x
to those generated byy
, and returny
.This is a way of chaining multiple feedback items together.
- Example:
incorrect() ; end()
- mark the student’s answer as incorrect, then end marking.apply(note1) ; apply(note2)
- apply feedback generated bynote1
, then feedback generated bynote2
.
- apply(feedback)#
If
feedback
is the name of a marking note, apply its feedback items to this note.You can give more than one name, to apply feedback from several notes on after the other.
If
feedback
is a list of feedback items generated by a function such assubmit_part
, apply them to this note.- Examples:
apply(validNumber)
- add the feedback from the notevalidNumber
to this note.apply(a,b)
- add the feedback from the notea
to this note, and then add the feedback from the noteb
.apply([submit_part(gaps[0]["path"]), submit_part(gaps[1]["path"])])
- mark the first two gaps and add their feedback to this note.
- apply_marking_script(name, studentanswer, settings, marks)#
Apply the marking script with the given name, with the given values of the variables
studentanswer
andsettings
and withmarks
marks available.Any feedback items generated by the marking script are applied to this note.
The built-in marking scripts are stored in the marking_scripts folder of the Numbas source repository. Use the name of the script without the
.jme
extension as thename
parameter of this function.- Example:
apply_marking_script("numberentry",studentAnswer,settings+["minvalue":4,"maxvalue":5],1)
- mark this part using the number entry part’s marking script, but with the minimum and maximum accepted values set to 4 and 5.
- submit_part(path[, answer])#
Submit the part with the given path. If
answer
is given, the answer stored for that part is overwritten with the given value. Returns a dictionary of the following form:[ "answered": has the student given a valid answer to the part?, "credit": credit awarded for the part, "marks": number of marks awarded, "feedback": feedback items generated by the part's marking algorithm ]
Custom part types can’t depend on other parts being available. However, you might want to allow the question author to provide the path of another part, or do something with this part’s gaps or steps, whose paths are listed in
gaps
andsteps
.
- mark_part(path, studentanswer)#
Mark the part with the given path, using the given value for
studentanswer
.Returns a dictionary of the following form:
[ "valid": is the given answer a valid answer to the part?, "credit": credit awarded for the part, "marks": number of marks awarded, "feedback": feedback items generated by the part's marking algorithm, "states": a dictionary mapping the name of each marking note to a list of feedback items, "state_valid": a dictionary mapping the name of each marking note to a boolean representing whether that note failed, "values": a dictionary mapping the name of each marking note to its value ]
This function is most useful in a custom marking algorithm for a gap-fill part, when you want to reassign the student’s answers to each of the gaps. For example, in a part with two number entry gaps, you could ensure that the lowest answer is marked by the first gap, and the highest answer is marked by the second. This would allow the student to enter their answers in any order, and the question author to set the expected answer for the first and second gaps to the lowest and highest correct answers, respectively.
- concat_feedback(items, scale[, strip_messages])#
Apply the given list of feedback items (generated by
submit_part
ormark_part
) to this note, scaling the credit awarded byscale
.If
strip_messages
istrue
, then all messages are stripped from the feedback items, leaving only items which modify the credit awarded.- Example:
Mark gap 0, and award credit proportional to the number of marks available:
let(result,mark_part(gaps[0]["path"],studentanswer[0]), concat_feedback(result["feedback"], result["marks"]) )
Pre-submit tasks#
Some Numbas extensions provide functions to perform tasks that take a long time, such as fetching a resource from the internet, or running code submitted by the student.
When you have a long-running task depending on the student’s answer that must be completed before the marking algorithm can run, you can run it as a pre-submit task.
A marking algorithm’s pre-submit tasks are defined in a note called pre_submit
.
This note must produce a list of promise
values, each resolving to a dictionary of values.
All the usual marking algorithm variables are available to this note.
These values are then collected into a dict
value called pre_submit
, available to the rest of the marking algorithm.
The marking algorithm runs once all of the tasks have finished. In the mean time, the part shows no feedback.
Tasks should be deterministic: they are assumed to produce the same results for identical student answers. The results of tasks are cached, so the tasks will only be run when the student’s answer changes.
When an attempt at an exam is resumed, the cached results of any pre-submit tasks are used, so the tasks don’t have to run again.
Example#
This example demonstrates the mechanics of pre-submit tasks by performing the otherwise-pointless task “wait for a while”.
The student’s answer will determine the number of seconds to wait.
First, define a JavaScript function wait(time)
which performs the task of waiting:
var promise = new Promise(function(resolve, reject) {
setTimeout(function() {
resolve({
seconds_waited: new Numbas.jme.types.TNum(time)
})
}, time*1000);
});
return new Numbas.jme.types.TPromise(promise);
Then add the pre_submit
note to the marking algorithm of a number entry part:
pre_submit:
[ wait(studentAnswer) ]
Next, add a feedback line to the mark
note to show the value returned by the task:
mark:
feedback("I waited {pre_submit['seconds_waited']} seconds before marking your answer.");
apply(base_mark)
When you run this question, note that the first time you submit a given number, there is a wait before the feedback is shown. If you submit the same answer again, the feedback is shown immediately, because the result of the task was cached.