Skip to content

Governments and Big Tech Unite to Set Rules for Responsible AI Amid Proliferation of Generative AI Models

As the global proliferation of artificial intelligence (AI) and generative AI (GenAI) models accelerates, governments and big tech firms are collaborating to establish responsible AI regulations. The United States, G7 nations, and other countries are introducing AI guidelines and executive orders.

Governments worldwide are intensifying their efforts to regulate artificial intelligence, especially the rapid growth of generative AI models.

Global Regulatory Landscape:

United States Executive Order
US President Joe Biden issued a new executive order on AI, urging developers of powerful AI systems to share safety test results and vital information with the US government. The order sets standards for AI models trained on massive computational clusters. The US government emphasizes international cooperation, engaging with countries like India to understand AI governance frameworks.

G7 Guiding Principles and Code of Conduct
The Group of Seven (G7) nations introduced guiding principles and a voluntary code of conduct for AI developers, focusing on ethical AI practices. These principles aim to prevent AI applications that undermine democratic values, harm individuals or communities, or pose substantial risks.

UK AI Safety Summit
The United Kingdom hosted an AI Safety Summit to address long-term risks associated with AI technologies. The UK government established the "Frontier AI Taskforce," collaborating with major AI companies to assess risks and access AI models. However, there is reluctance to set up a global AI regulator.

Many countries, including Canada, the US, China, Brazil, and Japan, have initiated AI regulatory measures. These efforts include drafting laws, guidelines, and frameworks for governing AI.

India, as a founding member and Council Chair of GPAI, is actively participating in international discussions on AI governance and responsible AI development.

The US executive order calls for companies developing foundation models to notify the government when training them and to share red-team safety test results. This directive targets large language models created by tech giants like Microsoft, Google, Meta, and Hugging Face.

The impact of the executive order on global businesses is unclear, especially concerning AI solutions built on APIs from US-based foundation models and large language models. The challenge lies in ensuring compatibility with the order's intent to protect US interests.

The executive order mandates the National Institute of Standards and Technology (NIST) to set AI standards for critical infrastructure sectors. These standards are aimed at preventing the misuse of AI models for engineering biological materials, which could impact areas like drug discovery.

To combat the spread of fake news, the order requires the development of content authentication and watermarking guidance for AI-generated content.

The US government calls for actions to counter adversaries' military use of AI, addressing concerns about AI weaponization. Autonomous Weapon Systems (AWS) are becoming a global focus in military technology.

The executive order urges Congress to pass data privacy legislation to protect Americans, particularly children.

Safety programs for AI in healthcare and support for AI-enabled educational tools are also part of the executive order.

The AI community remains divided on the regulation of foundation models, large language models, and artificial general intelligence (AGI). While some experts raise concerns about the risks associated with AI's exponential growth, others emphasize the benefits and argue that AI is far from becoming sentient.

As governments and big tech firms collaborate to establish AI regulations, the landscape of responsible AI development is evolving. The global implications of these initiatives on international businesses, especially those leveraging AI models, APIs, and solutions, will require continuous monitoring and adaptation. The debate on AI regulation will persist as AI technologies continue to advance.

As governments and the tech industry unite to shape the future of AI, responsible development and governance become paramount. The path forward involves a balancing act between innovation, ethical concerns, and international collaboration to ensure the responsible advancement of AI technologies.