Advertisement

Canada’s proposed AI law too vague, Big Tech executives say

Click to play video: 'Implications of AI Deepfakes'
Implications of AI Deepfakes
RELATED: Implications of AI Deepfakes – Jan 31, 2024

Representatives from Big Tech companies say a Liberal government bill that would begin regulating some artificial intelligence systems is too vague.

Amazon and Microsoft executives told MPs at the House of Commons industry committee today that Bill C-27 doesn’t differentiate enough between high- and low-risk AI systems.

The companies said abiding by the law as written would be costly.

Nicole Foster, director of global artificial intelligence and Canada public policy for Amazon, said using the same approach for all applications is “very impractical and could inadvertently stifle innovation.”

The use of AI by a peace officer is considered high-impact in all cases, she said — even when an officer is using auto-correct to fill out a ticket for a traffic violation.

Story continues below advertisement

 

Click to play video: 'Health Matters: Artificial intelligence in healthcare'
Health Matters: Artificial intelligence in healthcare

“Laws and regulations must clearly differentiate between high-risk applications and those that pose little or no risk. This is a core principle we have to get right,” Foster said.

Breaking news from Canada and around the world sent to your email, as it happens.

“We should be very careful about imposing regulatory burdens on low-risk AI applications that can potentially provide much-needed productivity boosts to Canadian companies both big and small.”

Microsoft gave its own example of how the law doesn’t seem to differentiate based on the level of risk that particular AI systems introduce.

An AI system used to approve a person’s mortgage and handle sensitive details about their finances would be considered the same as one that is used to optimize package delivery routes using public data.

Industry Minister François-Philippe Champagne has been offering some information about amendments the government expects to put forward to the bill to ensure it is up-to-date.

Story continues below advertisement

But in spite of that additional detail, companies said the definitions in the bill are still too ambiguous.

Click to play video: 'Navigating dating trends in the age of AI'
Navigating dating trends in the age of AI

Amanda Craig, senior director of public policy at Microsoft’s office of responsible AI, said not differentiating between the two would “spread thinly the time, money, talent and resources of Canadian businesses — and potentially mean finite resources are not sufficiently focused on the highest risk.”

Bill C-27 was tabled in 2022 to target what are described as “high-impact” AI systems.

But generative AI systems such as ChatGPT, which can create text, images and videos, became widely available to the public only after the bill was first introduced.

The Liberals now say they will amend the legislation to introduce new rules, including requiring companies behind such systems to take steps to ensure the content they create is identifiable as AI-generated.

Story continues below advertisement

Earlier this week, Yoshua Bengio, dubbed a “godfather” of AI, told the same committee that Ottawa should put a law in place immediately, even if that legislation is not perfect.

Bengio, the scientific director at Mila, the Quebec AI Institute, said a “superhuman” intelligence that is as smart as a human being could arrive as soon as in a few years.

Advanced systems could ultimately be used for cyberattacks, he said, and the law needs to get out ahead of that risk.

AI already poses risks. Deepfake videos, which are generated to make it look like a real person is doing or saying something that they never did, can be used to spread disinformation, said Bengio.

Sponsored content

AdChoices