AI Governance - a Hiroshima AI Processes Framework Approach
- Emmanuel Iserameiya
- Jun 16, 2024
- 3 min read
Updated: Dec 2, 2024
In continuation of my ongoing research on AI Governance frameworks, below is a summary of AI Governance from a Hiroshima AI Processes Framework Approach.

Introduction
The Hiroshima AI Processes framework, developed through the Hiroshima AI community's and international stakeholders' collaborative efforts, provides comprehensive guidelines for the ethical and responsible development and deployment of AI systems. This framework integrates ethical considerations, risk management, and regulatory compliance into AI processes, emphasising a proactive approach to AI governance. The goal is to align AI technologies with societal values and expectations, ensuring their safe, secure, and trustworthy use.
Key Principles (a summary)
Ethical Considerations
Respect for Human Rights: AI systems should enhance human well-being and avoid causing harm.
Fairness and Non-Discrimination: AI development must include principles of fairness, transparency, and respect for human rights, overseen by ethical review boards.
Ethical Impact Assessments: Conduct assessments to evaluate the potential impacts of AI systems.
Risk Management
Structured Risk Management: Organisations must identify, assess, and mitigate risks related to privacy, security, and societal impacts throughout the AI lifecycle.
Regular Risk Assessments: Implement continuous monitoring and update risk assessment processes regularly.
Transparency and Accountability
Documentation and Communication: Organisations should ensure stakeholders understand how AI systems operate by documenting and communicating decision-making processes.
Accountability Mechanisms: Establishing clear lines of responsibility and processes for redress.
Inclusivity and Stakeholder Engagement
Diverse Stakeholder Inclusion: Engaging with communities, users, and impacted individuals to incorporate diverse viewpoints into AI development.
Trust and Acceptance: Ensuring AI systems are more equitable and better suited to meet the needs of all users.
Continuous Improvement and Adaptability
Ongoing Monitoring: Establishing processes for continuously evaluating and updating AI systems.
Staying Informed: Keeping abreast of new developments in AI ethics, governance, and technology.
Implementation Requirements
Establishment of an AI Management System
AI Governance Framework: Developing policies, procedures, and controls that govern AI activities.
Integration with Organisational Strategies: Aligning AI risk management processes with broader organisational goals.
Continuous Monitoring and Improvement
Regular Reviews and Updates: Conducting ongoing evaluations and implementing corrective actions based on lessons learned.
Adaptation to Technological Advancements: Ensuring AI practices evolve with new technological developments.
Training and Awareness
Staff Training: Providing appropriate training for employees involved in AI activities.
Awareness Programs: Keeping staff informed about the latest developments and best practices in AI management.
Stakeholder Engagement
Effective Communication: Engaging with stakeholders to ensure transparency and accountability.
Incorporating Feedback: Using stakeholder input to drive continuous improvement.
Incident Management and Reporting
Incident Response Mechanisms: Establishing procedures for identifying, reporting, and addressing AI-related incidents.
Transparency in Incident Management: Ensuring accountability and transparency in managing incidents.
Third-Party Management
Compliance with AI Management System: Ensuring third parties involved in AI activities adhere to the organisation's AI management framework.
Monitoring Third-Party Performance: Conducting due diligence and establishing contractual obligations for third parties.
The Hiroshima AI Processes framework provides a holistic approach to managing AI technologies, emphasising governance, transparency, fairness, and resilience. By adhering to these principles and requirements, organisations can ensure AI technologies' ethical, responsible, and effective use. The framework's comprehensive guidelines help navigate the complexities of AI governance, enhancing trust in AI systems and promoting their ethical deployment. Implementing the Hiroshima AI Processes requires a proactive approach and a commitment to continuous improvement, ultimately contributing to the sustainable and ethical growth of AI applications.
Get in touch;
If you have any questions regarding the above,
If you have been tasked with implementing an AI Governance framework and have questions on how, or
If you want to learn more about AI governance and its tentacled risk management requirements in general.
References
Hiroshima AI Community. (2021). Hiroshima AI Processes: Ethical and Responsible AI Guidelines. Retrieved from [Hiroshima AI Community](https://hiroshima-ai-community.org)
National Institute of Standards and Technology (NIST). (2023). Artificial Intelligence Risk Management Framework (AI RMF 1.0). Retrieved from [NIST AI RMF](https://doi.org/10.6028/NIST.AI.100-1)
Hagendorff, T. (2020). The ethics of AI ethics: An evaluation of guidelines. Minds and Machines, 30(1), 99-120.
Comments