You are here: Home » Technology » News
Business Standard

Like human beings, man-made artificial intelligence also has a dark side

Though AI technology is perceived to be the next big thing, it also seems to have implications that go beyond its generic notion of making life easier for humans

BS Web Team  |  New Delhi 

Artificial intelligence

(AI) is one of the emerging concepts shaping the world of today. From analysing big data to powering autonomous vehicles, this has created new avenues for humans to experiment and explore. Though the AI is perceived to be the next big thing in technology, it also seems to have implications that go beyond the generic notion of it making life easier for humans.

A research by Stanford and has revealed that a agent that was meant to form street maps from aerial images was found to be hiding information in a clandestine manner to use it later stealthily. This incident came to light when researchers were working on improving the process of turning satellite imagery into Maps’ Street View feature, according to a news report in and start-up news portal the Tech Crunch.

ALSO READ: shuts down AI system after creating its own language

This is not the first time that an AI-based program has been found to learn new tricks to improvise its work beyond the purpose it was designed for. Last year, AI Research Lab (FAIR) also had to shut down chatbots powered by the social media giant’s AI-engine after they appeared to have been conversing in a language that only they understood.

The two examples above show that AI has, along with other things, also imbibed a dark side from humans. Though it is designed to work on data sourced from humans, its basic nature to learn, adapt and improvise makes it change its course from what it is designed for to what it is capable of — that can be horrendous and scary.

The AI is also seen to have inherited a sense of bias from humans. In 2016, an investigation by ProPublica, a US-based non-profit organisation, found that an AI-based software (COMPAS) used by judges in some US states to calculate risk score of a person re-committing a crime again, was biased against people of colour.

Such biases can also be seen in today’s smartphones with AI-driven beauty mode, which brightens the frame and considers white tone as beautiful. Maybe this is how the program is designed to work, but there is no denying that the technology has biases, just as humans do.

First Published: Thu, January 03 2019. 12:33 IST