It is important to keep diversity in mind when developing new AI solutions, it is a question that pops up at almost every AI conference, but we are still a long way away from answering that enquiry definitively.
Alessio Bagnaresi, the Microsoft sales lead for AI solution, said "Artificial intelligence is a technology that promises to revolutionise a variety of industries from healthcare, education, agriculture and transportation, especially in respect of driverless cars. Our aim is to use this technology to change industries and also really try to improve the society".
As the Redmond company sees it, AI should be used to amplify human ingenuity rather than replace it all together.
Take Saqib Shaikh for instance. Shaikh is a blind software engineer at Microsoft that used the firm’s cognitive AI solutions to create an app that relays information about the world to him. This functionality even extends to reading the facial expressions of his colleagues. Take a look at the video below or try out Microsoft’s SeeingAI app for iOS yourself.
This app allows Shaikh to engage with the world in a more meaningful way, even though he is blind. As you heard in the video it started with computers being able to talk and that has evolved into something infinitely more useful thanks to his ingenuity.
How Microsoft develops AI
In the not-so-distant future, what will happen if an autonomous vehicle controlled by artificial intelligence collides with a human, who is to blame?
Humans use AI therefore the human is responsible for the faults that might be caused by AI. Sadly, it is not as cut and dry as that. To explain the need to look at how Microsoft develops AI, or rather the key pillars it follows when developing solutions for its customers.
The first pillar is fairness. Apologies to future robot friends, but in 2018 AI is still very dumb when it enters this world. It still needs to learn a lot and while it can learn quickly one must be aware of what is being taught. When programming AI, one needs to look at every single angle and while the bot is operating it needs to be monitored constantly to ensure that biases are accounted for. Programming with a bias will result in a biased output.
On the top of this, AI must be inclusive, for the sake of the people who will use it and the sake of the firm creating the solution. "There are close on one billion people with disabilities and we need to provide equal opportunities to everybody," explains Bagnaresi. "We want to use this technology to empower people to do more," she continues.
Other key considerations when Microsoft develops AI solutions include reliability and safety. Privacy and security are rather important as well. The anonymisation of data is vital in a world where so much personal information is available and for its part, Microsoft is re-engineering all of its software to comply with global regulation such as GDPR. It’s also an ethically sound practice that feeds into one of the more important pillars of development – transparency.
Because AI is able to gather and learn from thousands upon thousands of data points easily, it’s vital to inform customers about what exactly the AI is doing with all of that information. More and more people are starting to understand the value of their data and want it to be protected. Facebook is learning this the hard way as it tries to make its users safe by investing in information protection but this has had the knock-on effect of losing the firm money and investor confidence.
And finally, one of the questions at hand – who is accountable for AI?
One would hope that we are closer to answering that question following the accident involving an autonomous Uber vehicle earlier this year but we really aren't. "That is a topic of big debate," says Bagnaresi answering the question about who is responsible for an accident involving a self-driving vehicle. "That is one of the reasons these technologies are not yet commercially available. That is an area that needs to be regulated with laws before anything enters a production environment," she adds.
To put it into simple terms, the laws that will govern AI will likely come from policy makers which is a bit concerning given some US Senators don’t know how Facebook makes money. Microsoft is also working with firms such as IBM and Google in a bid to understand why AI makes the decisions it does.
The underlying message within Bagnaresi’s talk is that diversity is the key to developing AI. No one person can account for every bias in the world and that’s okay but a firm must consider the impact its AI solution could have on everybody.