Skip to content

Feds set rules on use of AI in government services amid wider testing

OTTAWA — The department that oversees the federal social safety net has quietly started testing artificial-intelligence systems that could one day make it faster and easier to get answers about benefits and services.
15806319_web1_190305-RDA-Canada-Government-AI-PIC
File photo by THE CANADIAN PRESS Employment and Social Development Canada Chief Data Officer Sandy Kyriakatos poses for a photo in Gatineau, Quebec. The federal government is setting the ground rules for how departments and agencies can use artificial intelligence to make decisions about benefits and services.

OTTAWA — The department that oversees the federal social safety net has quietly started testing artificial-intelligence systems that could one day make it faster and easier to get answers about benefits and services.

A small team inside Employment and Social Development Canada is experimenting with ways to simplify navigating one of the largest service organizations in the country, handling public pensions, employment insurance, family benefits and disability supports. The department has a mix of offices, call centres and correspondence centres.

It’s hoping to have people directly interact with bots instead of humans, including in online chats for people seeking information about government programs.

An early draft of the department’s artificial-intelligence strategy, obtained by The Canadian Press under access-to-information laws, suggests the risks of chatbots, in particular, include “providing incorrect information to Canadians,” “producing incoherent content,” “or the reproduction of undesirable behaviour.”

The worst-case scenario? Internally, a chatbot could tell an employee to “explore a ‘catastrophic’ action” — think a Terminator-like order to kill all humans — while externally a bot could start replicating hate speech. A Microsoft-created Twitter bot did that three years ago after interacting with enough users and “learning” to mimic things they wrote. The decline was driven by people deliberately messing with the bot but it took only one day.

The strategy notes that ESDC needs to manage legal risks, ethical questions and logistical issues, not to mention “public perception” — and rapid technological advances that mean “the answers to these risks are moving targets.”

“We have a higher standard. We must meet a higher standard. Like it or not, I can decide not to go buy stuff (at one store) and go (to another). That’s not an option with government,” Sandy Kyriakatos, the department’s chief data officer, said in a recent interview.

On Monday, the government set the ground rules for how departments and agencies can use artificial intelligence to make decisions about benefits and services, or finding new uses for the technology for long-term projects like preserving and teaching Indigenous languages.

Talking at a conference of government workers Monday morning — hours before she resigned from cabinet — Treasury Board President Jane Philpott said anyone who thinks the era of artificial intelligence is just on the horizon is mistaken and Canadians are ready to get answers from machines about government services.

“When people go online and do their Christmas shopping, AI is influencing how they do their Christmas shopping,” she said after her morning talk. “We in government want to make sure we take advantage of the same kinds of tools to provide good services for people, but be extremely open in how it’s done.”

In setting new rules, Philpott said departments will have to be able to explain why a decision was made on a particular file, beyond saying it was up to a computer.