A resource is a single Numbas exam, which students access via the VLE.

Resources are automatically created when you launch a new Numbas activity from the VLE as an instructor. You must select a Numbas exam to use, and then any students who launch the same activity will be shown the exam.

Creating a new resource

Form with options to upload a Numbas package, or select a ready-made exam.

The first thing you see on creating a new Numbas resource.

Either upload an exam package that you have downloaded from the Numbas editor or, if any editor links have been created, select an exam from the list.

When you download an exam package from the Numbas editor, you must use the SCORM package option.

Once you’ve selected an exam, you will be shown the dashboard for the resource.



The dashboard screen.

When you open a resource as an instructor, you are first shown the dashboard. This view offers a few actions to do with the resource, and the navigation bar at the top has links to other views.

Report scores back to VLE

The LTI provider can automatically report scores back to the VLE.

Click the Report scores back to VLE button to begin this process. This may take some time; you’ll be shown either a success message or any errors encountered while reporting scores.

This isn’t supported by every VLE.

Discount question parts

Discounting a question part removes it from the score calculations: any marks students have earned for that part are discounted.

You might want to do this if an error is found in a question.

To discount a part, click on the Discount a question part button on the dashboard.

You are shown a list of all the question parts in the exam. Click a Discount this part button next to a part to discount it. You can choose whether to remove the part from the total available for the exam, or to award everyone full marks for the part. These have different effects on the weighting of other parts in the exam - removing a part from the total will increase the weighting of other parts, while awarding full marks will increase everyone’s total score. Think carefully about what you want to do.

If you discount a gapfill part, all of its gaps are discounted. If you discount an individual gap, the other gaps are unaffected.

List of question parts. Part a has been discounted.

Discounting a question part.

Download scores as CSV

Click this button to download a .csv file containing the scores for each student who has attempt the resource.

The columns of the file are:

  • First name

  • Last name

  • Email address

  • Username

  • Percentage score

The values in the name, email address and username fields come from the VLE. The username field might not correspond exactly to the student’s username on the VLE; in particular, Blackboard preprends usernames with cuid:.

The file might take a while to produce; it’ll be available in the Reports tab once it’s ready.

Student progress

From the dashboard, click View individual student progress and grant access tokens to view the Student progress table.

The Student progress table lists the names of students who have attempted the activity, along with their scores as calculated following the grading method, and the number of attempts they have made.

You can narrow down the displayed list by entering a name in the Search for a student box.

Note that only students who have launched the activity are listed - the LTI provider has no way of knowing about students who have access to the activity through the VLE but have never launched it.

Access tokens

When the number of attempts students are allowed to make is limited, circumstances can arise in which you want to allow particular students another attempt.

To do this, click the plus symbol in the Access tokens column next to the student’s name in the Student progress table.

To revoke an access token, click the minus symbol next to the student’s name in the Access tokens column. Attempts started with an access token that is subsequently removed will not be deleted.

When the student launches the activity, they will be offered the opportunity to start a new attempt.

Validate a receipt code

If enabled (see Email attempt receipts to students on completion?), students are emailed a receipt on completion of an attempt. The receipt contains information about their attempt, such as time and score, as well as a code that you can use to confirm that a receipt is valid.

On the Validate a receipt code page, copy a code from a student into a box.

If the code is valid, you’ll be shown the information from the receipt, as well as a link to view the attempt’s data in more detail.


A validated receipt code.

If the code is invalid, or corresponds to an attempt at a different resource, you’ll be told so.


An invalid receipt code.


Click on the Attempts button at the top of the page to view the attempt management screen.


The attempt management screen.

Download attempts summary as CSV

Click the Download attempts summary as CSV button to obtain a .csv file with information on every attempt at this activity.

The columns of the file are:

  • First name

  • Last name

  • Email address

  • Username

  • Start time, in YYYY-MM-DD HH:MM:SS.ffffff+HH:MM format.

  • Completed? (Either completed or incomplete)

  • Total score

  • Percentage (total score as a percentage of marks available)

  • One column giving the total score for each question

The file might take a while to produce; it’ll be available in the Reports tab once it’s ready.

Download all attempt data as JSON

Click the Download all attempt data as JSON button to obtain a .json file containing all data to do with attempts on this resource, including part and question scores, all SCORM data, completion status, and the values of question variables.

The file might take a while to produce; it’ll be available in the Reports tab once it’s ready.

Review an attempt

Click the Review button to view a student’s attempt as they saw it.

This is useful when a student queries the mark they were awarded for a part of the exam.

Note that review mode always opens attempts as if they were completed, even if the student has not yet ended the exam.

Remark an attempt

Click the Remark button to manually change the score awarded for a question part.

You are shown a list of every question part in the exam. Click the pencil icon on the row corresponding to the part you want to change, and enter the new score. The new score is saved as you type, and the totals for the question and the whole exam are recalculated automatically.

List of question parts. Part a gap 0 of question 1 has been discounted, and question 2 part a gap 0 has had its score manually set to 3.

Remarking an attempt.


After updating the exam package associated with a resource, you can automatically remark all attempts using the Remark tab.


By clicking on the Data button, you can see a timeline describing actions the student took during their attempt.

Items shown include moving between questions, submitting answers, and marks awarded. This page will update in real-time as the student continues their attempt.

Click the Download attempt data in JSON format button to obtain a .json file containing all data pertaining to this attempt, including question and part scores, completion status, and the values of question variables.

Numbas uses the SCORM standard to store data about attempts. By clicking on the View raw SCORM data for this attempt button, you can see all of the SCORM data model elements stored for a particular attempt.

This is most useful for debugging connection errors, to confirm that data has been saved.

If Most recent value only is ticked, only the most recent value for each element is shown. Untick it to see every value that the element has taken since the start of the attempt.

You can type a regular expression in the Search for an element box to narrow down the displayed list of elements.

Delete an attempt

Click the Delete button to delete an attempt. This is permanent; the student will be able to start a new attempt next time they launch the activity.

Reopen an attempt

Sometimes students accidentally close their attempts before they mean to. Click the Reopen button to allow a student to complete their attempt. The next time that they launch the activity, they will be able to resume the attempt as if they had only paused it.

Beware that the standard Numbas settings allow a student to see the correct answers to every question once they have finished their attempt. If you’re concerned about this, it’s often better to make the student start a new attempt, rather than reopen the previous one.


The Statistics page shows some statistics derived from attempts at this resource.

Attempt completion

The Attempt completion table shows the completion status of attempts. Don’t worry if some attempts are still marked as “incomplete” after the deadline has passed - scores for incomplete attempts are still counted.

An attempt will have the status “Not attempted” if the student opened the resource, but did not click the “Start exam” button. This can happen if the student’s device has a problem which causes the exam not to load, or if it does load but they just don’t start!

Summary statistics

The Summary statistics table shows the mean, median and quartiles for some statistics about the resource. The Time taken row is a rough measure of how long the student spent in their attempt; see Time spent for more information. Be careful when using this measurement: if a student completes their attempt in several sittings, the whole intervening time will be included, and even when a student has an attempt open, they might not be giving it their attention for the whole time.

Next is a breakdown of scores at each question, as a bar chart. The attempts at each question are classified as “not attempted”, “incorrect” (score 0), “partially correct”, or “correct” (the maximum score for the question).

Score distribution

The distribution of scores for the whole exam and for each question are shown as decreasing plots. The horizontal axis represents percentage score, and the vertical axis represents the proportion of attempts achieving at least that score.

Attempt times

Each attempt at the resource is shown as a span of time, with a dot at the recorded start time and another at the recorded end time.

Time spent

The distribution of time spent at attempts is shown as a decreasing plot. The horizontal axis represents lengths of time, and the vertical axis represents the proportion of attempts at least that long.

The ‘time spent’ is a rough measure of how long the student spent in their attempt. This time is increased each time attempt data is saved, based on the time since their current session started. It is not simply the difference between the start and end times: if the student closes their attempt for a long time and then resumes it later, the time that it was closed is not counted.

Be careful when interpreting this measurement: students might leave an attempt open while they do something else, so there are often large outliers.



The resource settings screen.

Replace exam package

If you discover an error in your exam, you can update it by downloading it again from the editor and clicking the Replace exam package button.

Any new attempts will use the latest version of the exam package. Because the new version might have changed in a way that is incompatible with existing attempts, for example by removing or rearranging question parts, any attempts started with the old package will by default continue to use the old package. If you know that the new package is compatible with the old one, for example if you’ve just corrected some text or fixed a bug in some code rather than changing the structure of the exam, tick Make existing attempts use this version. All attempts using the old package will be updated to use the new one.

If the new version of the exam fixes problems with marking, you can try automatically remarking existing attempts.

Grading method

Specify how a student’s score for the activity is calculated.

  • “Highest score” will use the highest total score from any of the student’s attempts.

  • “Last attempt” will use the total score from the attempt which the student began last.

Include incomplete attempts in grading?

If ticked, incomplete attempts will be included when calculating the student’s score for the activity.

It’s normally good to leave this on, so that students who forget to click the End Exam button won’t be penalised.

Maximum attempts per user

How many attempts at the resource can each user take?

If set to 0, then there is no limit.

You can grant extra attempts to individual students with access tokens.

When to show scores to students

When a student reopens an activity, they are shown a summary of their attempts. You might not want to immediately show students their scores on this screen.

  • “Always” means the student will see scores for all attempts, including incomplete attempts.

  • “When attempt is complete” means the student will only see their score for an attempt once it is complete.

  • “When review is allowed” means the student will only see their score after the date specified in the Allow students to review attempts from setting.

  • “Never” means that no scores are shown to the student, even after they’ve completed their attempt.


This only controls the display of scores by the LTI provider. If you want to hide scores from the students, you must also turn off the score feedback options in the exam editor.

When to report scores back

Specify when students’ scores are reported back to the consumer. Some VLEs make reported scores available to students immediately, which you may not want.

  • “Immediately” - scores are reported as soon as they change, i.e. whenever a student submits an answer.

  • “On completion” - a student’s score is reported when they complete an attempt.

  • “Manually, by instructor” - Scores are only reported when an instructor clicks the Report scores back to VLE button on the dashboard.

Allow students to review attempts from

Specify when students are allowed to re-enter completed attempts in review mode.

If left blank, students can review their attempts at any time.

If a date and time are set, students may only review their attempts after that time.

Instructors may always review students’ attempts, from the Attempts tab.

Available from and available until

Specify when students are allowed to access the resource.

You can set none, one or both of the options Available from and Available until.

If Available from is before Available until, the resource will be available only between those two times. Use this for a resource which should only be available for a certain period, such as an exam.

If Available from is after Available until, the resource will be unavailable between those two times, but available any time before Available until and any time after Available from. Use this for a resource which should be made unavailable for a certain period, such as a bank of practice material which students should not have access to during a summative assessment period.

Any students who have the resource open when it becomes unavailable will be forced to quit.

Email attempt receipts to students on completion?

If ticked, then when a student completes an attempt at this resource they will be emailed a receipt summarising their attempt.

The receipt contains a code which instructors can use to confirm the receipt’s contents. See Validate a receipt code.

Require a lockdown app?

You can require that students access this resource through the Numbas lockdown app or, if configured, Safe Exam Browser. See the documentation on lockdown-apps.

If you select Numbas lockdown app, you can give a password that the student must submit to launch the resource, or leave the field empty to use the LTI provider’s default password.

If you select Safe Exam Browser, you must select a settings file to use; these must be installed by a server administrator.

Whichever app you use, you can choose whether or not to show the password to the student before the app launches.

When a student launches a resource requiring a lockdown app, they will be shown a button to launch the app, along with a link to install it if they haven’t already.

Screenshot of the "Launch in the Numbas lockdown app" screen. The resource's name and context above a button to install the app, and a button to launch in the Numbas lockdown app.

This is what a student sees when they launch a resource which requires the Numbas lockdown app.


The remark resource view.

Remarking a resource.

The Remark tab provides an interface for automatically re-running attempts at the resource using the latest version of the exam package. If the total score awarded for an attempt is different to the saved score, you can overwrite it.

When the page loads, you are shown all attempts at the resource. You can remark individual attempts, or remark all the attempts automatically.


Remarking an exam is computationally intensive, so your browser may appear unresponsive.

If Use unsubmitted answers is ticked, then any answers entered by the student but not submitted will be considered to be submitted. You can use this in cases where a student forgot to submit their answers and you want to have them marked anyway.

After you click Remark all attempts, each attempt will be remarked in turn. A progress bar shows the proportion of attempts that have been remarked, and an estimated time until completion.

To stop the process of remarking all attempts, click Stop marking. If an attempt is currently being remarked, it can’t be interrupted - the process will stop after that attempt is finished.

You can choose to Show all attempts, only attempts with changed scores, or only attempts with either increased or decreased scores.

When one or more attempts have been remarked and produced different scores, the Save all changed attempts button becomes available. After clicking this button, any changed data corresponding to the shown attempts is saved to the database. This data will appear in the timeline for the attempt and will affect the reported scores for the attempt.

Each attempt has its own Remark button, which will remark only that attempt. When an attempt has been remarked, the change in total score is shown. If the total score is different to that stored in the database, a Save button appears. Clicking the button causes changed data corresponding to the attempt to be saved to the database, updating its timeline and reported score.

Access changes

Access changes are a means of changing deadlines or allowing a different number of attempts at the resource to certain students.

Each access change applies to a list of students.

To create an access change, click on the Access changes tab, then Add an access change.

There are several fields that you can modify. Leave a field entirely blank to keep the standard value from the resource’s settings.

In the Description field, describe what the access change is for and who it applies to, such as “25% extra time”, or “Late submission for Elliot D”.

Access changes take effect immediately. Any students who are completing an attempt at the time that their access changes will receive the new details immediately. If the resource becomes unavailable for them, the attempt will end immediately. Otherwise, a message will appear on their screen describing the new deadline.

Availability dates

The fields in the Availability dates section change the time period in which the affected students can complete attempts at the resource. You can either set new fixed start and end dates, or fill in the Extend the deadline by field to extend the resource’s normal Available from date.

If several access changes apply to a student, the changes from the last created access change are used.

Exam duration

If the exam associated with the resource has a duration set, then you can extend it, either by an absolute length of time or by a percentage of the standard duration.

If several access changes apply to a student, the change in the last created access change is used.

Number of attempts

The Maximum attempts per user field overrides the resource’s Maximum attempts per user field.

The student is allowed whichever is the greatest of the resource’s normal maximum number of attempts and the numbers specified by any access changes affecting the student.

A value of zero in this field does not mean “no change”, it means that the affected students are allowed to start as many attempts as they like.

Lockdown app

The Require a lockdown app? field override the resource’s Require a lockdown app? field.

The settings behave as follows:


The resource’s settings will be used.


Students affected by this access change will not be required to use a lockdown app.


Students affected by this access change will be required to use the Numbas lockdown app. You can change the password with the Password for the Numbas lockdown app field. If you leave it empty, the password defined by the resource’s settings will be used.

Safe Exam Browser

Students affected by this access change will be required to use the Numbas lockdown app. You can change the SEB settings file to use with the SEB settings field.

Applies to

Specify which students the access change applies to by giving lists of usernames or email addresses. These are compared with the usernames and email addresses sent by the LTI consumer.


Note that the LTI consumer can choose what it send for these fields, so they might not match the values you use in other systems. The access change form shows the username and email address it received from the LTI consumer for your account, as a guide.


When you ask to download one of the resource-level reports, it’s compiled in the background. Once it’s ready, you can download it from the Reports tab.

Reports are automatically deleted after a fixed period of time set by the administrator; the default is 30 days.

Test run

Click the Test run button to launch the Numbas exam. Data will not be saved - this feature is solely a convenience for instructors to check the contents of the exam.