Thursday, 3 April 2025
24.9 C
Singapore
26.8 C
Thailand
20.7 C
Indonesia
27 C
Philippines

Chinese tech companies face higher costs due to new EU AI rules

New EU AI rules, effective August 1, will raise compliance costs for Chinese tech firms, affecting innovation. Learn the implications.

The European Union (EU) is set to implement the worldโ€™s first comprehensive artificial intelligence (AI) regulations on August 1. These new rules are expected to significantly increase compliance and assessment costs for Chinese tech companies operating within the EUโ€™s 27 member states. Industry experts have highlighted the challenges of these regulations, especially in terms of innovation.

New rules and their implications

The Artificial Intelligence Act was approved by the EU Council in May, following its passage by the European Parliament in March. The Act aims to protect fundamental rights, democracy, the rule of law, and environmental sustainability from high-risk AI applications. At the same time, it seeks to boost innovation and establish Europe as a leader in AI technology.

Some Chinese AI firms, such as Hong Kong-based Dayta AI, are already preparing for the financial impact of these regulations. Patrick Tu, co-founder and chief executive of Dayta AI, predicts compliance and assessment requirements will increase the companyโ€™s research and development (R&D) and testing costs by 20 to 40 percent. This increase will cover additional documentation, audits, and technological measures.

Balancing regulation and innovation

The introduction of the AI Act reflects a global push to establish AI regulations amidst the rise of generative AI (GenAI) services. GenAI refers to algorithms that create new content, such as audio, code, images, text, and videos, in response to short prompts. Despite concerns about overregulation, Tanguy Van Overstraeten, a partner at Linklaters and head of the law firmโ€™s technology, media, and telecommunications (TMT) group in Brussels, believes the EUโ€™s goal is to create an environment of trust.

The AI Act categorises AI technology based on potential risks and impacts. It covers prohibited practices, high-risk systems, transparency obligations, governance, post-market monitoring, information sharing, and market surveillance. The regulation also requires member states to establish regulatory sandboxes for real-world testing, which allow companies to test AI applications within set boundaries for up to 12 months.

Non-compliance with certain AI practices can lead to administrative fines of up to 35 million euros (US$38 million) or up to 7 percent of the offending firmโ€™s total worldwide annual turnover, whichever is higher.

Comparing global regulations

Dayta AIโ€™s Tu noted that the EUโ€™s focus on data quality will ultimately enhance the performance and fairness of AI solutions. He also compared the EUโ€™s user rights-focused approach with the regulations in China and Hong Kong, which he believes focus more on enabling technological progress and aligning with government priorities.

Chinaโ€™s GenAI regulations, implemented on August 15 last year, require AI service providers to adhere to core socialist values and avoid generating content that threatens national security or promotes terrorism, extremism, or other harmful ideologies. Alex Roberts, a partner at Linklaters in Shanghai, pointed out that these regulations can confuse multinational corporations that are unfamiliar with such requirements.

Roberts also mentioned that Chinaโ€™s AI regulations are more state-led, whereas the EUโ€™s regulations focus on user rights. Despite these differences, he believes the core principles of both regulatory frameworks are similar, including transparency, data protection, accountability, and providing clear guidance on AI products.

The State Council, Chinaโ€™s cabinet, has listed a comprehensive AI law in its legislation plans for 2023 and 2024, though a draft law has yet to be proposed. Other Asian countries, like South Korea, are also working on AI regulations. South Koreaโ€™s draft โ€œAct on Promotion of AI Industry and Framework for Establishing Trustworthy AIโ€ is still under review.

Roberts concluded that governments in the Asia-Pacific region increasingly look to the EUโ€™s AI regulations as a model for their legislation. This trend allows businesses to advocate for more consistent and harmonised cross-market rules.

Hot this week

Uwant D500 Pro Wet and Dry Vacuum Cleaner review: Designed for comfort and ease of use

The Uwant D500 Pro Wet and Dry Vacuum Cleaner offers powerful suction, real-time sterilisation, and ergonomic design for effortless cleaning. With self-cleaning, AI-powered drying, and intuitive controls, it ensures a seamless, low-maintenance experience for modern homes.

Roblox introduces new parental controls to enhance child safety

Roblox introduces new parental controls, allowing parents to block games, restrict friends, and monitor their childโ€™s activity for better safety.

Exabeam introduces Nova, an agentic AI that boosts cybersecurity operations

Exabeam unveils Nova, a proactive AI agent that boosts security team productivity and reduces incident investigation time by over 50%.

Fitbit users now have until 2026 to migrate to Google accounts

Fitbit users now have until February 2, 2026, to migrate their accounts to Google accounts or risk losing their data and service access.

Facebook introduces friends-only feed to cut out algorithmic content

Facebookโ€™s new Friends tab removes algorithmic recommendations, letting you see only posts from friends. It is now rolling out in the US and Canada.

Qualcomm expands AI research with MovianAI acquisition

Qualcomm has acquired Vietnamese AI research firm MovianAI to boost its AI development in smartphones, PCs, and software-defined vehicles.

Roblox introduces new parental controls to enhance child safety

Roblox introduces new parental controls, allowing parents to block games, restrict friends, and monitor their childโ€™s activity for better safety.

Anthropic introduces Claude for Education, a new AI chatbot plan for universities

Anthropic launches Claude for Education, an AI chatbot plan for universities that offers advanced learning tools and administration support.

Exabeam introduces Nova, an agentic AI that boosts cybersecurity operations

Exabeam unveils Nova, a proactive AI agent that boosts security team productivity and reduces incident investigation time by over 50%.

Related Articles