Development Phase Activity Booklet (text only)
Published 5 June 2025
These materials were produced by The Alan Turing Institute through an extended engagement with personnel at the UK Ministry of Justice.Â
Ethics ProcessÂ
Introduction to the SAFE-D FrameworkÂ
- Read the Introduction Booklet and familiarise yourself with the Project Lifecycle Model.Â
SAFE-D Principles BookletÂ
- Read the SAFE-D Principles Booklet and familiarise yourself with the principles and their core attributes. Â
Design Phase Activity BookletÂ
- SAFE-D Identification Workshop ExerciseÂ
- Litmus TestÂ
- Stakeholder Engagement WorksheetÂ
- Additional activities where relevantÂ
Development Phase Activity BookletÂ
- SAFE-D Reflection Workshop ExerciseÂ
- Development Phase Questionnaire Â
- Stakeholder Engagement WorksheetÂ
- Additional activities where relevantÂ
Deployment Phase Activity BookletÂ
- SAFE-D Assurance Workshop ExerciseÂ
- Deployment Phase QuestionnaireÂ
- Stakeholder Engagement WorksheetÂ
- Additional activities where relevantÂ
Model Development PhaseÂ
Model DevelopmentÂ
Covers technical tasks such as training, testing, and validating the machine learning model to ensure it’s suitable for its purpose.Â
Lower-level lifecycle stages and their ethical significance
Model Selection & TrainingÂ
This stage involves choosing one or more algorithms for training a model. Factors influencing this decision include:Â
- Access to computational resourcesÂ
- Predictive performance of the modelÂ
- Characteristics of the dataÂ
Ethical significance:Â
- While there are technical reasons for responsibly selecting and training a model, a key aspect is the model’s interpretability and explainability.Â
- Generally, more complex models (like convolutional neural networks) are harder to interpret compared to simpler ones (like linear regression).Â
- The choice of algorithm should align with the model’s intended use and the system it will be part of.Â
Model Testing & ValidationÂ
Model testing and validation involves evaluating a model using various metrics, including its accuracy on new data that wasn’t part of the original training set.Â
Ethical significance:Â
- When a dataset is divided into training and testing sets (internal validation) or when a model is assessed with completely new data (external validation), there are opportunities to evaluate more than just performance.Â
- Testing a model’s generalisability to new contexts helps ensure it is both sustainable and fair, meaning it should maintain similar accuracy and performance levels even when validated externally.Â
Establishing Monitoring TestsÂ
Effective monitoring of a model’s performance in its runtime environment relies on setting up metrics to check if the model is functioning within its intended parameters.Â
Ethical significance:Â
- While tests typically measure accuracy and global interpretability, it’s also important to assess system-level performance, such as efficiency and resource usage.Â
- The team responsible for the model will need their expertise for this stage, along with input from systems engineers, software engineers, and end users.Â
- This stage is crucial for establishing collective responsibility throughout the project’s lifecycle. It occurs at an important point in the project and relies on clear and accessible communication among all team members.Â
Model ¶Ù´Ç³¦³Ü³¾±ð²Ô³Ù²¹³Ù¾±´Ç²ÔÌý
This task involves documenting both formal and informal aspects of the model and its development process, including:Â
- Data sources and summary statisticsÂ
- The model used Â
- Evaluation metricsÂ
Ethical significance:Â
Clear and accessible documentation is vital for responsible project governance for several reasons:Â
- It ensures results can be reproduced and supports values like public accessibility in open researchÂ
- It promotes accountability and transparency in decision-makingÂ
- It helps individuals affected by the technology seek redress for any harms that may occur during its design, development, or deploymentÂ
Development Phase SAFE-D ReflectionÂ
Workshop Exercise: SAFE-D Development CheckpointÂ
Activity OverviewÂ
Now that you’re in the Project Development Phase, it’s important to reflect on the SAFE-D Principles and assess how your project aligns with them. Use the following steps and prompts to guide your discussion and documentation.Â
³§³Ù±ð±è²õ:Ìý
1) Review previous workÂ
- Look back at the answers and action strategies you developed during the Design Phase workshop activities to refresh your memory Â
2) Use the Miro boardÂ
- Track any changes, progress, and new questions that have arisen in the Development PhaseÂ
- You can create a new section on the board, or add post-its in a different colour to your original answersÂ
3) Evaluate SAFE-D goals & objectivesÂ
- Revisit the objectives you set during the SAFE-D Specification exercise. Assess whether these strategies are working effectively or if they need adjustmentsÂ
Prompt questions for discussion:Â
- What changes have occurred since the Design Phase that impact the SAFE-D Principles?Â
- How has the project progressed in relation to the ethical principles? Are there any areas where you feel you have made significant progress?Â
- Have any new questions or concerns emerged regarding the SAFE-D Principles as the project develops?Â
- Are the strategies you implemented in the Design Phase proving effective? What challenges have you encountered?Â
- Have you received any feedback from stakeholders that may influence your assessment of the SAFE-D Principles?Â
¶Ù´Ç³¦³Ü³¾±ð²Ô³Ù²¹³Ù¾±´Ç²ÔÌý
Make sure to document your reflections and discussions in a clear format. This will not only provide a record of your progress but also facilitate ongoing communication among team members and stakeholders.Â
Importance of ReflectionÂ
Reflecting on the SAFE-D Principles at this stage is crucial for maintaining ethical integrity throughout the project. By actively assessing your project’s alignment with these principles, you can identify potential risks and ensure that ethical considerations are prioritised as you move forward.Â
Development Phase QuestionnaireÂ
The Development Phase Questionnaire is the next formal checkpoint in the ethics process. Â
³§³Ù±ð±è²õ:Ìý
Use previous insightsÂ
- Leverage the insights and discussions from your SAFE-D workshops and Litmus Test to inform your responses in the assessmentÂ
Assign team membersÂ
- Consider nominating different team members to collaborate on various sections of the assessment. This allows for diverse input and thorough exploration of each principleÂ
Complete the questionnaireÂ
- The Development Phase Questionnaire consists of a series of questions organised around the SAFE-D principles. Each principle has 3-4 questions that require a simple Yes/No answer, along with justification or evidenceÂ
Justification and evidenceÂ
- Ensure that for each answer, you provide sufficient justification or evidence. This could include references to workshop insights, project documentation, or stakeholder feedbackÂ
Track changes and decisionsÂ
- Remember that the purpose of this assessment is not to hinder innovation but to help identify and mitigate risks early in the project lifecycleÂ
Document the assessmentÂ
- After completing the assessment, compile and document the responses. This will serve as a useful reference for future stages of the project and facilitate transparency with stakeholdersÂ
³§³Ü²õ³Ù²¹¾±²Ô²¹²ú¾±±ô¾±³Ù²âÌý
2.1a: Is the use case free from any risks to fundamental rights and freedoms? [Safety]Â
2.1b: Are there sufficient security measures and access controls in place? [Security]Â
2.1c: Have you tested the model for robustness to data drifts and adversarial attacks? Please answer N/A if this is a one-off analysis or will not be used with new data [Robustness]Â
2.1d: Are the outputs of the model consistent and aligned to expectations? [Reliability]Â
2.1e: Is the performance within acceptable thresholds? [Accuracy & performance]Â
´¡³¦³¦´Ç³Ü²Ô³Ù²¹²ú¾±±ô¾±³Ù²âÌý
2.2a: Are the escalation pathways and approval pathways clear for this project? i.e. who is approving this model? [Answerability]Â
2.2b: Are the roles and responsibilities of the project team and wider stakeholders clearly defined? [Answerability]Â
2.2c: Have you liaised with the data owner(s) on any limitations, quality issues or bias in the data set? [Clear data provenance & lineage]Â
2.2d: Do you know who the data owners are and have clear data documentation of all features? [Clear data provenance & lineage]Â
2.2e: Is the documentation on the model accessible and easy to understand? [Accessibility]Â
2.2f: Have you checked that any engineering or selected features are appropriate with the relevant domain experts? [Auditability]Â
2.2g: Have you signed off the performance and error rates with the appropriate stakeholder(s)? [Auditability]Â
2.2h: Are all decisions made documented with justifications? (For example, any data input methods and why they were chosen) [Auditability]Â
2.2i: Are all relevant actions and analyses traceable and reproducible? [Traceability]Â
¹ó²¹¾±°ù²Ô±ð²õ²õÌý
2.3a: Is the project data collection strategy returning a representative sample of the population? [Non-discrimination]Â
2.3b: Is there a sufficient sample in any subgroup of interest for this analysis? [Non-discrimination]Â
2.3c: Have you considered including diverse voices and opinions in the design and development process, e.g. through engaging with civil society and affected communities? [Diversity & inclusivity]Â
2.3d: Have any features added by developers that could be affected by personal judgement been documented and mitigated? [Bias mitigation]Â
2.3e: Have you identified and mitigated potential proxies in your data, i.e., features in your data that may be associated with unjustifiable sources of differences in outcome, e.g., postcode with race, income with gender? [Bias mitigation]Â
2.3f: Has the model been calibrated across different metrics for each group and mechanism? [Bias mitigation]Â
2.3g: Is the model free of all potential feedback loops that may worsen existing inequalities? [Equality]Â
·¡³æ±è±ô²¹¾±²Ô²¹²ú¾±±ô¾±³Ù²âÌý
2.4a: Have you considered all the stakeholders that might need an explanation, either global (model level) or local explanation (individual level)? [Implementation & user training]Â
2.4b: Are the users sufficiently informed and knowledgeable to understand the explanation? [Implementation & user training]Â
2.4c: Can you clearly communicate how each feature was calculated or collected? [Interpretability]Â
2.4d: Are you able to create clear global and local explanations for each prediction? [Interpretability]Â
2.4e: Has the explainability of the model remained consistent as the size of the data set and the number of features has increased? [Interpretability]Â
2.4f: Where a complex model is selected, is the rationale documented as to why a simpler model cannot be used? [Responsible model selection]Â
2.4g: Is the explanation accessible and easy to understand for each stakeholder? [Accessible rationale explanation]Â
2.4h: Is the explanation reproducible? [Reproducibility]Â
Data ResponsibilityÂ
2.5a: If using historical data, are all data sets being used timely and applicable for deriving future insights? [Timeliness & recency]Â
2.5b: Have you ensured adherence to existing laws and regulations, including data protection law, privacy standards, and public sector duties? [Legal & organisational compliance]Â
2.5c: Have all the potential risks, assumptions, and limitations of this use-case been reviewed and signed-off? [Legal & organisational compliance]Â
2.5d: Are any of the recorded features free from human judgement? If no, have they been recorded and mitigated? Source integrity & measurement accuracy]Â
2.5e: Are there clear policies in place for data storage, sharing, documentation, and processing? [Responsible data management]Â
2.5f: If any data sets are linked and shared, have data protection principles been considered? [Responsible data management]Â
2.5g: Do the available data set(s) have the adequate sample size, representativeness, and contextual relevance for the use case at hand? If not, are the limitations documented? [Adequacy of quality & quantity]Â