Skip to Main Content

Course Design Blueprint: AI resilience in assessment

Understanding AI in assessment

When designing assessments for your course, it is important to consider the vulnerability and resilience of each assessment type in relation to the use of generative AI. A useful tool for the evaluation of your assessment is the AI Risk Measure Scale (ARMS), developed by the University of Greenwich (De Vita and Brown, 2023). 

The AI Resilience Measure Scale
AI Risk Measure Scale (ARMS) Description Examples
1 Very low Highly unlikely AI can be used to produce this type of assessment.

Assignments that embed authenticity in the design (e.g., field trip + reflective report), assignments that allow establishing the identity of the person (e.g., presentation, in-person exams). 

Subjective assignments that require personal reflection or creative thinking, such as personal narratives or artistic projects. These types of assignments are typically based on the student's opinions and insights, which are difficult to replicate using AI. 

2 Low

Potential use of AI to produce the assessment. 

Very unlikely to have a significant impact on quality and / or originality.

Assignments that draw on unique teaching material (e.g., novel cases produced by the tutor.

Assignments that have clear guidelines, such as solving math problems or coding exercises, where AI could assist but the student's approach or or solution is what is being evaluated as the main focus of the assignment.

3 Moderate

Moderate likelihood of AI use to produce the assessment. 

Could have moderate impact on quality and / or originality.

Assignments where AI could be used to assist students in completing the assignment, but the final work would still require the student's critical thinking, analysis and interpretation. 

Assignments that require a more complex analysis of a topic, e.g., critical analysis essay or scientific report. Students may use AI tools to help with data analysis, visualisation or interpretation in some areas. but the writing and argumentation are largely based on the student's understanding and critical thinking. 

4 High

Easy to use AI to produce the assessment. 

Could have significant impact on quality and / or originality.

Assignments that focus on well-published company case studies (e.g., Innocent, Apple. Boohoo, Starbucks etc.) and rather generic topics (e.g., advantages and disadvantages of FDI) which students can easily obtain through AI bots. 

Assignments that involve sophisticated algorithms or complex modelling such as financial forecasting, predictive analytics or image recognition, where students could use AI to generate both results and insights / commentary

5 Very High

Very easy to use AI to produce the assessment. 

Will have a significant impact on quality and / or originality.

Assignments that require students to produce summaries or abstracts of published articles, reports or research papers, this includes research proposals. These assignments require no input / modification from students and can be entirely produced by AI. 

Assignments that involve large scale data processing, such as machine learning projects or artificial intelligence simulations, where students could rely entirely on AI to create the work. 

CPD activity for AI and authentic assessment is provided through the Centre for Excellence in Learning and Teaching and can be booked through the CPD calendar.


 

Assessment and AI at Suffolk

We have undertaken an evaluation of AI vulnerability of the most commonly used assessment types at Suffolk, using ARMS. Key insights from the evaluation are:

  • AI risk is very much subject to interpretation, design and organisation of the assessment. 
  • The ever changing nature of AI technology means that assessments which are generally considered AI resilient have become more AI vulnerable in recent times, e.g., videos. 
  • Design and organisation of the assessment is crucial to increase AI resilience.
  • Authentic assessments will typically be more AI resilient by design. 
Summary of AI and Assessments UOS (2024)
AI Risk Assessment type
1 - very low risk In person exam (oral or written), OSCE / practical, practical assessment, class discussion, simulation, simulation, observation, research participation, physical artefact. 
2 - low Digital artefacts, reflection, job application / interview. 
2-3 low / moderate 4 - high Dissertation, report, portfolio, presentation, video, podcast, blog, discussion post, policy briefing, peer assessment. 
4 - high 5 - very high Essay, case study, research proposal, multiple choice test (online), book review

Next steps

Course teams have already been asked  to consider each of their assessments and their vulnerability or resilience, and where needed to complete a course modification. We strongly encourage course teams to reflect and evaluate the effectiveness of their chosen assessments as part of the marking and moderation process.