Academics Set to Shape New AI Code of Practice

In a significant development within the realm of artificial intelligence, a diverse group of academics has been tasked with drafting a Code of Practice for general-purpose AI (GPAI). This Code aims to clarify risk management and transparency requirements for various AI systems, including the widely recognized ChatGPT. The work of these academics comes at a crucial time as concerns over the ethical implications of AI technology grapple with the demands for innovation and safety.

The announcement of this academic-led initiative comes on the heels of questions raised by three influential Members of Parliament (MEPs) regarding the timing and international expertise of the appointed leaders. Despite these concerns, the working group comprises specialists from institutions around the world, ensuring a range of perspectives and expertise in the discussion.

At the helm of this initiative is Yoshua Bengio, noted for his pivotal role in the development of AI and often referred to as one of its “godfathers.” He will chair a group focused on technical risk mitigation, complemented by legal scholars and governance experts. Among them are law professor Alexander Peukert and AI governance authority Marietje Schaake, who bring unique insights that will guide the working group through the complexities of AI regulation.

The first draft of the Code is set to be released in November, following a workshop for GPAI providers scheduled for mid-October. This timeline is strategic, aiming to align with the broader context of the European Union’s AI Act, which will significantly depend on the forthcoming Code of Practice until formal standards are finalized by 2026. The urgency for this regulatory framework stems from the rapid advances in AI technology, which, while beneficial, pose significant risks if left unchecked.

What makes this initiative particularly vital is its focus on risk management and transparency. The AI systems in question not only impact businesses and governments but affect individuals in their everyday lives. For instance, AI chatbots like ChatGPT have demonstrated capabilities that raise questions about privacy, misinformation, and accountability. By developing a comprehensive Code of Practice, the group seeks to address these issues systematically, ensuring that AI technology remains safe, ethical, and beneficial for society.

Notably, the group’s composition reflects a thoughtful approach to the multifaceted nature of AI. As AI technologies increasingly influence social and economic governance, the necessity for interdisciplinary collaboration has never been more evident. Experts from technical, legal, and social spheres will come together to create guidelines that not only support technological advancement but also protect individual rights and broader societal interests.

The EU AI Act will serve as a cornerstone for this initiative. The Act outlines regulatory measures for high-risk AI, emphasizing the importance of safety and compliance for companies deploying such technologies. The Code of Practice will act as an essential supplement to the legislation, providing clarity on ambiguous areas that may hinder innovation while ensuring that stringent safety measures are in place.

The forthcoming first draft of the Code of Practice is expected to outline specific strategies for managing risk, including best practices for transparency and robustness in AI algorithms. Such details are crucial as stakeholders—ranging from tech giants to small startups—seek actionable insights into how they can comply with evolving regulations while maintaining their competitive advantage.

In conclusion, the development of this Code of Practice signifies a proactive stance taken by the academic community and policymakers to navigate the complex landscape of AI. By focusing on creating a framework that balances innovation with responsibility, this initiative promises to provide a roadmap for future AI developments that prioritize safety, transparency, and ethical governance. The impact of these efforts could shape the trajectory of AI technology and its integration into society for years to come.

Back To Top