Artificial intelligence has made it possible to automate decisions that human beings had been responsible for until now. Although many of these decisions are in the fields of entertainment and social media, automated decisions are also used in finance, education, the labour market, insurance companies, medicine, and justice. This phenomenon has far-reaching social consequences and brings up all sorts of questions, from “what will happen to the jobs that took care of making these decisions in the past?” to “how can we guarantee that those algorithms make fair decisions?”
Mary Bollender is a financially struggling single mother from Las Vegas. One morning in 2014, when her ten-year old daughter had a fever that would not go away, Mary decided to drive her to the emergency room. But when they got in the car, she found that it refused to start. The problem was not engine trouble or an empty tank: the bank had remotely activated a device that prevented her car from starting because her monthly car repayment was three days overdue. As soon as she paid, she would be able to get back on the road. Unfortunately, she did not have the money that morning.
As our society continues to become more complex, with data digitalisation expanding on a massive scale, cases like Mary Bollender’s will become increasingly common: automated decisions based on the systematic collection of personal data with potentially negative effects on our lives. Algorithms control which of our friends’ posts we see first on our Facebook news feed and recommend films we might want to watch on Netflix, but algorithms also decide whether a bank will give us a loan, whether someone awaiting trial can be released on bond, whether we deserve a postdoctoral grant, and whether we are worth of being hired by a company.
These computer programs or algorithms are not produced as a result of human programmers analysing and breaking down a problem and then feeding precise instructions into a computer. They are the fruit of complex mathematical operations carried out automatically that look for patterns in an ocean of digitized data. They are not like a recipe that sets out a list of ingredients and then tells us what do do with them step by step. It is more a case of “open the fridge, see what’s in there, rummage around the kitchen to find some pots and pans you can use, and then make me dinner for six.” These types of algorithms that are not explicitly designed by a programmer come under the concept of “machine learning”.
Machine learning algorithms are currently used to process data and determine whether a person will be able to repay a loan. The person who programs the algorithm compiles a database of people who have asked for loans in the past, including all kinds of data: the applicant’s sex and age, whether or not they repaid the full amount, whether and how often they were late with repayments, their average wage, how much tax they paid, in what city and neighbourhood they live, and so on. The algorithm applies a series of statistical formulas to all this data and then generates patterns that estimate the probability that a new potential client will repay a loan. Reliability is usually the only criteria used to develop these algorithms. Is the program accurate enough to replace a human being? If the answer is yes, it gets a green light.
It is not easy to determine reliability. Developing a truly reliable algorithm is partly a science and partly an art. When we think about software programs or artificial intelligence, we inevitably tend to anthropomorphise and imagine some kind of entity that follows a mental process similar to our own. But that is not how it works. An automated algorithm does not analyse the examples that we give it and then try to establish some kind of reasonable causal link between the data and the end result. The algorithm knows nothing about gender, age, financial conditions, unemployment, etc. It simply has a string of numbers and tries to find patterns that allow it to come up with the correct result as often as possible.
And this is where the problem comes in. A traditional program developed by a human being follows a certain logic, so it is possible to understand what it is doing. An automated algorithm is like a black box. We give it a certain input (the personal details of the person applying for a loan) and it gives us an output (the probability of that person repaying the loan). It is very difficult – or virtually impossible – to know why the program decided to reject or accept a particular loan application.
In the 1980s, the United States army commissioned scientists to develop an automatic image recognition program that could detect camouflaged tanks in real time. The scientists asked the military officers to provide numerous photos of different sites organised in pairs: one showing the site without a tank and the other showing the same site with a tank camouflaged among the greenery. The idea was to develop an algorithm that would come up with a series of criteria for detecting tanks. The program worked remarkably well.
It was one hundred percent accurate. In fact, it detected tanks that were so well camouflaged that a human being would not have noticed them. The researchers were so surprised that they decided to analyse what criteria the algorithm was using. After an in-depth study of the photographs and the algorithm, they realised that the program did not actually recognise tanks or anything remotely similar. Say the army had taken the photographs of the places without a tank at midday. The photographs with a camouflaged tank were taken at six in the afternoon. So in order to decide whether or not there was a tank in the image, all the algorithm had to do was look at the position of the sun.
Another example is self-driving cars. We like to imagine that cars driven by algorithms have some idea the meaning of a road, traffic lights, zebra crossings, a cyclist, another vehicle, and so on. But they are really just sophisticated versions of the tank-detecting algorithm. Their learning is highly contextualised and depends wholly on the response of the environment in which the algorithms are trained. And because they are black boxes, we will never know for sure how a self-driving car will react in a context that is significantly different to the one in which the algorithm was originally trained.
With enough training in very diverse contexts we can come up with algorithms that are truly robust and reliable. But there is a still more insidious problem: justice, or fairness. Given that we are talking about programs that recognise contextual patterns in a finite data set, and are not based on actual knowledge of the environment, no algorithm would ever consider reactivating Mary Bollender’s car so that she could drive her daughter to the hospital. It just knows who meets repayments and who doesn’t. Poor neighbourhoods have a much higher debt loss ration. A large percentage of single mothers tend to make mortgage and loan repayments late. An automated algorithm would no doubt reject a loan application by a single mother in that poor neighbourhood. The decision would no doubt be statistically correct. But would it be fair? Do we want to live in a world in which crucial life decisions are based on context-dependent statistic patterns?
The machine learning algorithms that are being developed now need the humanities. We have to create infrastructures for cooperation between software engineers and humanists. We need to develop a common language that allows sociologists, anthropologists, philosophers and artists to understand the basic functioning of this new kind of software, and computer engineers to think about how to embed ethics, peaceful coexistence, justice, and solidarity into the new software. Perhaps one day we will manage to develop the artificial superintelligence that Elon Musk is so keen on. But for now we should be much more worried about how existing programs can augment the racist, xenophobic and sexist biases that exist in our society.
Ramon Sangüesa | 14 March 2017
Buen post!.
Gracias por remarcar estos aspectos del desarrollo de algoritmos de IA. Es un problema que no queda restringido a la IA solamente, claro está, sino a todos los ámbitos donde unse ha programado un algoritmo que toma decisiones por o sobre nosotros.
Dentro de la comunidad tecnológica hay un grupo bien consciente de estos problemas y que colabora con otras disciplinas de ámbitos no tecnológicos: Fairness And Acountability and Transparency in Machine Learning (FAT-ML http://www.fatml.org/). Se centra, sobre todo, en como garantizar la equidad en los algoritmos de aprendizaje automático que, como sabes, son un parte importante de los algoritmos de IA, El año pasado co-organizamos la conferencia en NYC y espero acercarla aquí el 2018.
En cierta medida lo que comentas también es una preocupación de los participantes en la reunión de la semana pasada en Barcelona que dió lugar a la Declaración de Barcelona para el desarrollo y uso correcto de la Inteligencia Artificial: http://www.lavanguardia.com/ciencia/fisica-espacio/20170309/42678964251/barcelona-manifiesto-bases-inteligencia-artificial-europa.html
Saludos!
Ramon Crehuet | 15 March 2017
És un article molt interessant però crec que culpabilitzes massa la manca de comprensió de l’algorisme com a causant de la manca de justícia. La frase de “els diners no tenen principis” l’he sentida més d’un cop, i no la deia un algorisme. I em temo que, per una mare soltera en un barri empobrit, aconseguir un crèdit ja era difícil abans del Machine Learning.
De la mateixa manera, has demostrat que sí que es pot saber en què es basa un algorisme de Machine Learning quan has explicat que van poder entendre que només la posició del sol determinava el resultat d’on hi havia un tanc i on no. Cal, és clar, la voluntat i l’esforç d’entendre-ho, que potser sovint no hi és. Per cert, l’eficàcia del predictor no era de 100%! N’hi havia prou en donar-li una foto sense tanc presa a les sis de la tarda. L’error va ser clarament humà.
I és que, en general, els criteris que volem que segueixi l’algorisme els triem nosaltres. Però potser en un moment diferent del qual l’algorisme s’aplica, i això en difumina la responsabilitat. Les coses que no vam tenir en compte en el moment de disseny ens xocaran quan veiem la decisió injusta de l’algorisme. Per tant, calen articles com aquests que ens ho recordin.
David Casacuberta | 16 March 2017
Gràcies pels vostres comentaris, Ramons.
No coneixia la iniciativa FAT-ML, que trobo molt necessària. Queda apuntada.
És cert que la manca de transparència de l’algorisme no és ni molt menys la única font d’injusticia. Ni tan sols la principal. Només cal pensar en com està dissenyat l’algorisme darrere el negoci d’Uber. Simplement volia destacar una font extra d’injustícia de la que normalment no som conscients.
Efectivament l’algorisme no era ni molt menys 100% efectiu. Era un algorisme ben dolent en realitat. El que volia indicar és que, quan van fer la primera prova, amb les primeres fotos amb i sense tancs, la impressió general era que tenia un 100% d’efectivitat perquè localitzava tots els tancs, per molt amagats que estiguessin.
Leave a comment