The UN called on Wednesday to impose a moratorium on certain systems ofartificial intelligence like facial recognition, the time to put in place safeguards to protect human rights.
“Artificial intelligence technologies can have negative or even catastrophic effects if used without taking sufficient account of how they affect human rights,” said Michelle Bachelet, High Commissioner for Human Rights of ONU.
She called for an assessment of the risks presented by different systems that rely on artificial intelligence for the right to privacy or freedom of movement and expression and then to ban, or in any case strongly regulate those which present the greatest dangers.
But while waiting for these assessments to be carried out, “States should impose a moratorium on technologies that potentially present a great risk”, underlined the former Chilean president, during the presentation of a new report of her services. devoted to this topic.
In particular, she cited as an example the technologies that allow automatic decision-making or those that establish profiles.
“Harming human lives”
“AI systems are used to determine who can benefit from public services, decide who has a chance of being hired for a job and, of course, influence what information people can see and share in line, ”she stressed.
This report, which was commissioned by the Human Rights Council – the UN’s highest body in this field – looked at how these technologies have often been put in place without how they are working or their impact has not been properly assessed.
AI malfunctions have prevented people from receiving welfare, finding employment, or leading to arrests of innocent people based on poorly trained facial recognition systems unable to properly recognize people with physical features. Africans for example.
“The risk of discrimination linked to decisions based on artificial intelligence – decisions that can change, stigmatize or harm human life – is all too real,” Ms. Bachelet insisted.
The report pointed out that these artificial intelligences are trained with the help of huge databases, which are often built in an opaque way.
These databases themselves can be poorly made, outdated and therefore contribute to human rights violations.
In particular, the report highlights the increasing use of AI-based systems by law enforcement agencies, including predictive methods.
When AI uses biased databases, this is reflected in the predictions and tends to affect areas wrongly identified as high risk.
Real-time and remote facial recognition is also increasingly used around the world, which can lead to permanent location of people.
“We cannot afford to continue to try to catch up with the bandwagon when it comes to AI and allow it to be used with little or no control and to repair the human rights consequences after the fact,” he said. insisted the High Commissioner, even if she recognizes that “the power to serve people of AI is undeniable”.
“But so is the ability of AI to fuel human rights violations on a gigantic scale and in almost invisible ways,” she warned.
“We must act to put human rights-based safeguards on AI for the benefit of everyone,” she stressed.