Google recently announced that its discussions with European Union (EU) regulators have been a success. The talk was focused on the EU’s groundbreaking artificial intelligence (AI) regulations that could very shape our digital future.
Thomas Kurian, the head of Google’s cloud computing division, revealed the ongoing talks and emphasized the importance of developing AI technologies safely and responsibly. Addressing concerns raised by the EU, Google is working on tools to tackle issues such as the potential difficulty in differentiating between human-generated and AI-generated content.
Google’s Proactive Stance
Kurian highlighted the inherent risks associated with AI technologies but also stressed their tremendous potential to create significant value for individuals. In response to the EU’s worries, Google is focusing on ensuring that humans can accurately distinguish content generated by AI. As part of this effort, the company recently introduced a “watermarking” solution that enables the identification of AI-generated images.
These developments underscore the proactive stance of Google and other major tech companies in driving private sector control of AI. Google presents itself as a gatekeeper of information integrity, even before formal regulations are put in place.
AI systems, such as ChatGPT and Stability Diffusion, are rapidly evolving and pushing the boundaries of what our current technology can achieve. Increasingly, computer programmers are utilizing tools like ChatGPT to assist them in tasks like code generation.
EU’s Copyright Concerns
However, a key concern for EU policymakers and regulators is the potential mass production of content based on copyright-infringing material using generative AI models. This raises the risk of detrimental consequences for artists and creators who rely on royalties to make a living. Generative AI models are trained on large sets of publicly available internet data, much of which is protected by copyright.
To address these concerns, the European Parliament recently passed legislation known as the EU AI Act, which aims to ensure oversight of AI deployment in the EU. The act includes provisions aimed at preventing the violation of copyright laws in the training data used for generative AI tools.
Google’s Kurian acknowledged the importance of understanding and addressing these concerns. The company is actively collaborating with EU authorities to ensure that the worries presented by regulators are taken into account. Google recognizes AI as a major battleground, with companies vying for leadership in its development, particularly in generative AI.
The ability of generative AI to churn out new content based on user input, such as music lyrics or code, has captivated both researchers and business leaders. However, the rapid advancement of AI has also raised concerns about job displacement, misinformation, and bias.
Google itself has faced internal criticism and concerns raised by top researchers and employees regarding the company’s handling of AI development and ethics. This includes issues surrounding the introduction of Bard, Google’s generative AI chatbot, and its perceived hastiness and lack of ethical considerations.
Kurian emphasized Google’s willingness to embrace regulation and its commitment to collaborating with governments across the EU, the UK, and other countries. He believes that these powerful technologies require responsible regulation and is supportive of efforts to ensure their proper implementation.
While regulators are often criticized for their relatively slow response to emerging technologies, many companies, including Google, are taking proactive steps to establish their own frameworks for AI governance.
The UK, for instance, has introduced a framework of AI principles to empower regulators, allowing them to guide the responsible development and deployment of AI. In the United States, the administration of President Joe Biden and various government agencies have proposed frameworks for regulating AI as well.
The collective aim is to strike a balance between harnessing the potential of AI technologies and mitigating the associated risks, thereby ensuring the responsible and beneficial use of AI in society.