HomeAI News
Canada flags concern on AI-generated deepfake disinformation campaigns
2

Canada flags concern on AI-generated deepfake disinformation campaigns

Hayo News
Hayo News
November 18th, 2023

The Canadian Security Intelligence Service (CSIS), Canada’s primary national intelligence agency, raised concerns about the disinformation campaigns conducted across the internet using artificial intelligence (AI) deepfakes.

Canada sees the growing “realism of deepfakes” coupled with the “inability to recognize or detect them” as a potential threat to Canadians. In its report, CSIS cited instances where deepfakes were used to harm individuals.

“Deepfakes and other advanced AI technologies threaten democracy as certain actors seek to capitalize on uncertainty or perpetuate ‘facts’ based on synthetic and/or falsified information. This will be exacerbated further if governments are unable to ‘prove’ that their official content is real and factual.”

Since 2022, bad actors have used sophisticated deepfake videos to convince unwary crypto investors to willingly part with their funds. Musk’s warning against deepfakes came after a fabricated video of him surfaced on X (formerly Twitter) promoting a cryptocurrency platform with unrealistic returns.

The Canadian agency noted privacy violations, social manipulation and bias as some of the other concerns that AI brings to the table. The department urges governmental policies, directives and initiatives to evolve with the realism of deepfakes and synthetic media:

“If governments assess and address AI independently and at their typical speed, their interventions will quickly be rendered irrelevant.”

CSIS recommended a collaboration among partner governments, allies and industry experts to address the global distribution of legitimate information.

Canada’s intent to involve the allied nations in addressing AI concerns was cemented on Oct. 30, when the Group of Seven (G7) industrial countries agreed upon an AI code of conduct for developers.

As previously reported by Cointelegraph, the code has 11 points that aim to promote “safe, secure, and trustworthy AI worldwide” and help “seize” the benefits of AI while still addressing and troubleshooting the risks it poses.

Reprinted from ARIJIT SARKARView Original

Comments

no dataCoffee time! Feel free to comment