1. Internal and external examiners
Includes: Postgraduate students with teaching responsibilities; responsibilities of Head of School
All assessments for programmes and modules leading to a University award must involve both one or more internal examiners and one or more independent external examiners. In addition, all members of academic staff will act as internal examiners in the subjects in their School.
The School has a duty to ensure that any individuals acting as internal examiners who are not academic staff of the University, are fully briefed and have relevant experience and sufficient knowledge of their subject area to assess students' knowledge and skills. For more information, please consult the following:
Policy on occasional teachers
Postgraduate students with teaching responsibilities may also act as internal examiners. For more information, please consult the following.
Policy on University of Nottingham students who teach
In respect of any assessment carried out by postgraduate students with teaching responsibilities, Heads of School should ensure that:
- They are supervised by a named member of academic staff
- They are provided with training appropriate to their role in the assessment process
- Assessments that contribute to the final module/programme mark:
1. Are moderated by a member of academic staff
2. Are limited, as far as practicable, to those elements of assessment that do not contribute to the degree classification
- Such assessments which do not contribute to the final module/programme mark are second marked where appropriate on a sampling basis by a member of academic staff.
The University of Nottingham policy on the appointment and responsibilities of External Examiners can be found online. For more information, please consult the following:
Appointment and responsibilities of External Examiners
Students should be informed by their School when and how they will receive results and the various stages of the process of distributing results.
lndependent Assessors for Apprenticeship Programmes
Where a specific integrated apprenticeship programme’s End-Point Assessment plan makes references to other specified assessment roles within the marking and assessment team arrangements, the responsibilities noted below apply to all defined roles within that team.
On integrated apprenticeship programmes the End-Point Assessment serves as a final credit bearing module for the degree. It also constitutes the summative assessment for the apprenticeship award. Each specific End-Point Assessment plan outlines the final classification criteria for the apprenticeship award. Thus assessment of this module will need to meet the requirements of the published apprenticeship End-Point Assessment plan in addition to contributing to the degree award.
The Institute for Apprenticeships and Technical Education (IfATE) prescribes that the assessment of the End-Point Assessment is performed by suitably qualified Independent Assessors. These are appointed internally or externally and they will provide the grade for the overall outcome for the apprenticeship award. Dependent on the programme, the grade to contribute to the degree award may also be provided by the Independent Assessor/s, and thus may not be a member of University staff.
2. Marking criteria
Includes: role of School Boards; requirements of the School; Marking criteria examples; double-marking
It is recognised that there is a need for School Boards to conduct the assessment of students in a manner that is appropriate to individual disciplines and to the methods of assessment employed.
All Schools are required to have written marking criteria across the full range of marks available (0-100) which is published in School Handbooks. Marking criteria must include categories from 70% to 100% and 0% to 40%. Marking advice should be available to markers in relation to all forms of assessment used within the School/Department.
The School's marking criteria should be signposted in its programme specifications (Section C. Supplementary Regulations, 3. Assessment).
For examples of marking criteria, please consult the following:
In pursuit of assessment practices that are fair, valid and reliable the University recognises double-marking (preferably "blind" where the first mark is not made known to the second marker) as good practice for all assessments where appropriate.
For apprenticeship programmes, to meet requirements outlined by the Institute for Apprenticeships and Technical Education (IfATE), the criteria outlined in the specific End-Point Assessment plan for each apprenticeship standard should be followed. For more information please consult the following:
Institute for Apprenticeships and Technical Education
Permitted differences to local School policy on marking criteria may be required, for example providing acceptable tolerances in assessment word count.
3. Overview of assessment policies and procedures
Includes: assessment process
Quality assurance of assessment consists of several elements. These can be distinguished according to the different stages of the assessment process at which they operate:
- Pre-assessment activities (including guidance for students and markers about the assessment activity; scrutiny of the assessment activity by internal or external examiners; calibration of standards among markers)
- The assessment itself (including measures to assure the fairness of assessment conditions, such as invigilation, appropriate adjustments for students with support plans, and external review of exam papers)
- Marking (including use of clear standards, anonymity of marking, and double marking where appropriate)
- Internal Moderation (checking that standards have been applied consistently during marking)
- Marks adjustment (adjustment of marks in light of a flaw in the assessment to maintain standards across cohorts)
- Scrutiny by External Examiners
- Approval of marks by Exam Board
The appropriate combination of these elements of quality assurance will vary by discipline and assessment task, so some variation in practice between Schools is to be expected and not all elements will be appropriate to all assessments. Further information regarding moderation and marks adjustment can be found below.
Includes: role of external examiners; sample size; moderation assessment report form
Moderation is the process of checking that standards have been applied correctly and consistently during the marking of an assessment. It is distinct from marking (including double marking) and from marks adjustment. Unlike double marking, moderation should not result in a change to the mark awarded to just one student; any changes to marks following moderation should be applied systematically to all affected students taking the assessment (e.g. all students who answered a given question). Unlike marks adjustment, moderation examines the application of standards to a unit of assessment, and does not compare marks across different cohorts.
Moderation is also distinct from the scrutiny of assessment by External Examiners. External Examiners will not be involved in the marking process as their role is to ensure that policies and regulations have been appropriately applied, that the assessment process is fair and robust and to ensure that standards applied are consistent with subject benchmark statements and other institutions.
All assessments that contribute to the determination or classification of an award must have been subject to moderation. Moderation must have been conducted by someone other than the first marker.
An absolute minimum number of student submissions for each assessment must be reviewed by the moderator. That number is determined by the number of students registered to take the assessment as follows:
Submissions to be included in moderation sample
|Number of students registered to take assessment, N||Minimum number of submissions to be included in moderation sample, M|
|| N/20 (i.e. 5% of submissions)
Schools should decide for themselves the appropriate sample size for moderation for each assessment and it is expected that this will often exceed the minimum. This decision should be based on the nature of the assessment, its weighting (i.e. number of credits), the number of first markers and their level of experience.
Sampling for moderation should be random but should include at least one submission from each mark range (where they exist): 0-39, 40-49, 50-59, 60-69, 70-100.
Schools should have a clear and explicit written policy governing the moderation of marking which should be available to students (for example via School handbooks and School Community Moodle pages). All markers in a School should be clearly briefed on these processes each academic year.
These processes should specify:
- What proportion of work is sampled during moderation for each type of assessment (subject to the above minimum)
- Whether the sample must include work of certain kinds (e.g. whether special attention is paid to work marked as borderline)
- What should happen if a moderator believes that standards have not been applied correctly
Schools should have a uniform system for fully recording the process of moderation so that examiners can clearly understand how moderation has occurred and what consequences have arisen from it. These records should be available to External Examiners as soon as possible and be tabled at relevant Examination Boards. An example of such a system is provided below from the Faculty of Arts:
For apprenticeship programmes, moderation will be in line with School policy, unless there are alternative requirements specific to the End-Point Assessment module. Moderation of this module for will be undertaken by the appointed Independent Assessors or by internal staff who are independent of prior teaching. This will be determined and arranged locally.
5. Marks adjustment
Includes: role of external examiners; Report form for the use of marks adjustment
A School is expected to consider data regarding marks on modules for which it has responsibility (for example, median mark and variance for each module) to identify possible issues with the delivery and assessment of the module prior to the Exam Board. Normally this analysis would also be taken into account in the School’s regular review of modules.
Where there has been a clear flaw in the assessment process (e.g. mistakes on a question paper), Schools have a responsibility to ensure that this does not have a significant impact on the fairness or robustness of the final marks. Where such a flaw is identified Schools should develop a suitable adjustment to be applied to all marks and report this to the External Examiners.
Where no potential flaws have been identified please complete the summary report form:
Where potential flaws have been identified by markers, reported by students or may come to light because of comparison of marks profiles (e.g. comparison of median mark and variance) with other modules taken by the same cohort or with the same module taken by previous cohorts. Please complete the marks adjustment report for the affected module(s):
Review and sign off of the marks adjustment form (including nil returns) should take place at the Exam Board. Any approved adjustment and its rationale should then be communicated clearly to all students who took the module (normally via the module feedback report) and communicated to Quality and Student Management Systems Team (within Registry and Academic Affairs) for report to Quality and Standards Committee (QSC) using the forms above.
The module level information report must detail the rationale for marks adjustment; the method used; the method used to inform students; and any remedial measures implemented as a result of marks adjustment. These forms which should be appended to the summary form submitted. A separate report form should be used for each module where marks have been subject to adjustment. Schools are expected to submit an annual report even if there is a nil return. Where there is a nil return a single summary form should be used.
Reports should be submitted at the same time as the main Examination Board recommendations in June/July (Undergraduate programmes) and by November (Postgraduate Taught programmes).
6. Viva Voce examinations
Viva voce examinations may only be used as part of a documented assessment procedure and should not be used as part of the consideration of borderline candidates or in the final decision making process.
The only exception to this is where an external body requires that a Viva Voce examination is held as a separate assessment forming part of the decision-making process. In such a case the School must publish a written policy that must be included in the student handbook or otherwise drawn to the attention of students. It should address the following issues:
- The name of the body that requires the viva voce examination to take place
- The role of the viva in the assessment/classification process
- Who has the authority to call a student for a viva
- Selection criteria
- Whether attendance is compulsory or voluntary
- Notice given to a student
- Procedures followed and documents produced
- Whether a student can request a viva
Formal records of all such vivas should be kept.
Schools are encouraged to organise meetings for external examiners with representative groups of students as a mechanism for ensuring quality and standards control but this should not form part of the assessment process.
7. Penalties and legibility
8. Marking by numbers
Includes: anonymous marking
For formal written examinations the University operates a marking by numbers (or anonymous marking) system. For more information about the marking by number system, please consult the following:
Marking by number
Schools are encouraged to mark coursework anonymously where possible.
Schools should ensure that clear statements of the responsibilities of those involved in computation, checking and recording of assessment data exist.
Includes: degree classification; progression
For the purposes of classifying undergraduate degrees, marks will be rounded at the stages detailed under Degree Classification for each individual module.
The University convention on rounding of numeric marks for all awards is as follows:
Marks should be rounded at two stages only:
- When two or more unit marks are computed (using a weighting formula), the result should be rounded into a single integer module mark
- When the overall weighted average mark has been computed, it should be rounded into a single overall integer mark, before a classification is assigned.
Rounding means that any mark of x.5 and decimal fractions above, becomes the next highest integer e.g. 69.5 is rounded to 70, 59.5 to 60, and so on. Decimal fractions below x.5 are rounded to the next lowest integer - for example, 69.4 is rounded to 69. For the purposes of rounding, the mark should be truncated and only the first decimal place is used (e.g. 69.45 becomes 69.4, which is rounded to 69).
Following the rounding convention set out above, overall marks of 39.5, 49.5, 59.5, and 69.5 will be rounded to 40, 50, 60 & 70 respectively.
For the purposes of progression, overall average marks will be rounded to the nearest integer so that marks of 29.5, 39.5 and 49.5 will be rounded to 30, 40 and 50 respectively. Decimal fractions below x.5 are rounded to the next lowest integer - for example, 69.4 is rounded to 69. For the purposes of rounding, the mark should be truncated and only the first decimal place is used (e.g. 69.45 becomes 69.4, which is rounded to 69).
If you have any problems or queries relating to this page, please contact: