The complex landscape of artificial intelligence (AI) is increasingly relevant across all fields and nations. Since AI affects every state, its benefits and drawbacks must be addressed collectively.
The recent passage of the Framework Convention on Artificial Intelligence by the Council of Europe marks a significant turning point in the quest for effective and long-lasting control of AI. The document is a statement about how European states will harness AI, a relatively new and powerful technology, to empower humanity while limiting its abuses.
The March 2024 adoption of a US-led United Nations General Assembly resolution on seizing the opportunities of safe, secure, and trustworthy artificial intelligence systems for sustainable development also highlights the common ground the United States shares with Europe and democracies around the world. While the US differs in some important areas with Europe, the significant agreement is valuable. This convention provides a foundation for others to build on.
Although new technologies often appear benign at first glance, there are many ways AI, in particular, can negatively impact the values held dear by so many around the world. From facial recognition systems interfering with privacy to machine learning that amplifies injustice, the potential cost of AI’s misuse is high.
Until now, there was not an effort at global governance of AI. This fostered a string of national laws that lack the capacity to address the transnational nature of AI adequately. A system of fractured development silos creates a wide gap wherein technological advancement is prioritized over ethics.
There are a few AI regulatory developments around the world for AI management worthy of note. First, the US House of Representatives debated the Digital Services Oversight and Safety Act of 2022 (H.R.6796) and the Algorithmic Accountability Act of 2023 (H.R. 5628). The AI Disclosure Act of 2023 (H.R. 3831) was also introduced before the House of Representatives.
Canada saw the introduction of the Artificial Intelligence Data Act (AIDA). The European Parliament began discussing the EU Artificial Intelligence Act in 2023, which represents a shift toward “hard law.” China is also addressing the issue and undertook the Interim Administrative Measures for the Management of Generative AI Services, which was enacted in 2023.
Recently, the United States passed the 2024 National Defense Authorization Act, which included a five-year implementation plan to adopt AI applications to accelerate decision advantage for both business efficacy as well as warfighting capability.
China also released the Global AI Governance Initiative, “calling for all countries under the principles of extensive consultation, joint contribution, and shared benefits to enhance exchanges and cooperation, work together to prevent risks, and develop an AI governance framework based on broad consensus, to make AI technologies more secure, reliable, controllable, and equitable.”
The Council of Europe’s Framework Convention, discussed above, provides a comprehensive legal structure to protect human rights, democracy, and the rule of law, besides promoting innovation. This strikes a balance where the use of technology should not be limited but society must ensure that development is in the right direction—with most of the technologies promoting ethical usage.
The Convention offers a set of guidelines for the advancement of AI that include respecting human dignity and upholding the idea that one should not infringe upon the rights and liberties of others. There is a significant urge to pursue explainability, which requires an explanation for each decision an AI makes and highlights how these systems have the power to significantly impact people’s lives in fields like criminal justice and healthcare. Additionally, the Convention calls for fairness in the creation of AI by considering prejudice in learning systems.
The Convention clearly aspires to international cooperation, which is important because it is a central element of the liberal international order. Numerous states agreed to the treaty, realizing that artificial intelligence is not a local problem. Such collaboration is required since artificial intelligence affects all states, and its advantages and disadvantages can only be addressed jointly. The agreement’s openness to other parties is crucial because it enables people from different backgrounds and cultures to voice their thoughts, enhancing and leveling the playing field.
As AI advances, concerns will grow about the agreement’s direction, applicability, and resource allocation. It implies that ethically right deeds now might turn unethical as technology develops, highlighting the necessity of a progressive framework for a code of ethics.
Since the development of AI must be accompanied by its ethical application, the Convention on Artificial Intelligence represents an important accomplishment in European regulation of the technology. Therefore, this treaty serves as a guide for how human rights, accountability, and openness should be upheld when integrating AI into society. The journey has only just begun. However, the political will and spirit to act, observe, and adopt the proper approach to governance can serve to achieve the benefits of AI.
Nevertheless, it will probably be difficult to put these rules and policies into foolproof practice. It will certainly be a work in progress. The European Council is certainly providing a road map for managing global mutual technology advancement concerns.
Huma Rehman is Director of Research at the Islamabad Policy Institute (IPI) and a Defense & Foreign Affairs Analyst. The views expressed are her own.