Representatives from Big Tech companies say a Liberal government bill that would begin regulating some artificial intelligence systems is too vague.
Amazon and Microsoft executives told MPs at a House of Commons industry committee meeting Wednesday that Bill C-27 doesn’t differentiate enough between high- and low-risk AI systems.
The companies said abiding by the law as written would be costly.
Nicole Foster, director of global artificial intelligence and Canada public policy for Amazon, said using the same approach for all applications is “very impractical and could inadvertently stifle innovation.”
The use of AI by a peace officer is considered high-impact in all cases, she said – even when an officer is using auto-correct to fill out a ticket for a traffic violation.
“Laws and regulations must clearly differentiate between high-risk applications and those that pose little or no risk. This is a core principle we have to get right,” Foster said.
“We should be very careful about imposing regulatory burdens on low-risk AI applications that can potentially provide much-needed productivity boosts to Canadian companies both big and small.”
Microsoft gave its own example of how the law doesn’t seem to differentiate based on the level of risk that particular AI systems introduce.
An AI system used to approve a person’s mortgage and handle sensitive details about their finances would be considered the same as one that is used to optimize package delivery routes using public data.
Industry Minister Francois-Philippe Champagne has provided the text of the amendments the government expects to put forward to the bill to ensure it is up-to-date.
But in spite of that additional detail, companies said the definitions in the bill are still too ambiguous.
Amanda Craig, senior director of public policy at Microsoft’s office of responsible AI, said not differentiating between the two would “spread thinly the time, money, talent and resources of Canadian businesses – and potentially mean finite resources are not sufficiently focused on the highest risk.”
Bill C-27 was tabled in 2022 to target what are described as “high-impact” AI systems.
But generative AI systems such as ChatGPT, which can create text, images and videos, became widely available to the public only after the bill was first introduced.
The Liberals now say they will amend the legislation to introduce new rules, including requiring companies behind such systems to take steps to ensure the content they create is identifiable as AI-generated.
Craig said there are portions of the bill that go further than similar legislation in the European Union.
Rachel Curran, Meta’s head of public policy for Canada, said “AIDA has created a standard that doesn’t exist anywhere else in the world,” referring to the AI piece of Bill C-27, the Artificial Intelligence and Data Act.
Meta has the resources to meet it, she said, but warned “the compliance costs are incredibly high.”
“Would that mean that certain products may not be launched in Canada? Maybe.”
She said “the EU is setting a benchmark that is world-leading.”
“We should at the very least not exceed that.”
Meta also took issue with one of the government’s proposed amendments, which Curran said would classify content moderation or prioritization systems as high-impact. Meta owns the Facebook, Instagram and Threads platforms, which use algorithms to determine what content to show users.
“We respectfully disagree that these systems are inherently high-risk as defined in the legislation,” Curran said.
One of the criticisms of the Artificial Intelligence and Data Act has been a lack of consultation by the government before it was introduced.
In response to a question from Conservative MP Rick Perkins, the representatives of the four companies, some of the biggest in the AI space worldwide, said the government did not consult with them before tabling the bill.
“I find that shocking, frankly, that you wouldn’t have been consulted,” Perkins said.
Earlier this week, Yoshua Bengio, dubbed a “godfather” of AI, told the same committee that Ottawa should put a law in place immediately, even if that legislation is not perfect.
Bengio, the scientific director at Mila, the Quebec AI Institute, said a “superhuman” intelligence that is as smart as a human being could arrive as soon as in a few years.
Advanced systems could ultimately be used for cyberattacks, he said, and the law needs to get out ahead of that risk.
AI already poses risks. Deepfake videos, which are generated to make it look like a real person is doing or saying something that they never did, can be used to spread disinformation, said Bengio.