Is AI the future of humanity - Artificial intelligence and its limitations DW Documentary

Show video

Artificial intelligence, or AI, is considered a key technology for the future: Do you come to New York often? Mostly for work, but sometimes just for fun. It makes the work of doctors, psychologists, or police officers easier and is expected to make drivers or even curling players a thing of the past. In every aspect of everyday life, AI could help us make the best decisions: Should I move the rook or the bishop? Turn left, or right? Shoot or hold my fire? Date John or Jane? The relentless logic of algorithms is supposed to guarantee us a life free from errors. But lately, even programmers have been sounding the alarm. There’s a self-congratulatory feeling in the air.

If AI hasn’t lived up to its promise is it really as intelligent as it’s made out to be? Shit, this guy is really good. He’s not a guy, he’s a machine. What are they gonna do, replace us? And what are the limits of AI? A Turing machine like this, an early computer devised by the English mathematician Alan Turing, was the first machine capable of solving a puzzle more efficiently than the human brain. With its help, the British succeed in deciphering encrypted German radio messages at the height of World War 2, after countless military specialists had racked their brains in vain.

It marks the dawn of a new era with the development of devices that automate the work previously requiring human brainpower. At the beginning of automation, the goal was to reduce physical effort, that is, the amount of effort required. So, for centuries, a mill was considered an automated process. But over time, this approach would be applied to nonmaterial or mental work Nowadays we are dealing with a new form of automation, which we generally call artificial intelligence. In the 1950s, this development accelerated rapidly, with the promise of how Artificial Intelligence, AI, would optimize our lives. It was supposed to drive our cars, improve our education, provide us with healthy food, make the best medical diagnoses, and find just the right words to cheer us up.

Initially progress was slow. But that all changed in the early 2000’s, with new, powerful mainframe computers able to handle huge amounts of data. I was at Google, and I was at Google for a long time at that point. And then suddenly everyone in tech, everyone at Google, everyone, everywhere is like: Let's use AI to solve everything. AI is going to solve cancer.

AI is going to solve transportation. AI is going to solve education. I had somebody pitch me AI to detect genocide, right? And I'm like, what the F? Like Based on what? What are you training it with? What are the political stakes? No answers to this, right! People wanted to digitize all of reality, with the goal of knowing everything in real time.

There was something godlike about this idea, without taking into account that much of reality just can’t be reduced to zeroes and ones. Hi, Joanie, there is a beautiful rainbow outside. Gordon, Im programming You mean you turn the Poetry of the world into electrons? Fast forward to our time, and machines are now said to be capable of learning by themselves thanks to a completely different method of information processing and multi-layered training: so-called deep learning. There was a major advance in the deep learning systems based on their ability to read natural images so on the ImageNet challenge.

AlexNet won the ImageNet challenge sort of proved the efficacy of deep learning and that there was like a catalytic kind of AI goldrush. The ImageNet contest is an annual image recognition test for computer programs. For years, even the best of them got it wrong in every third guess, but in 2012 technology based on machine learning was suddenly able to bring the error rate down to 15%. Before this breakthrough, everything had to be explained to a program meant to recognize a face. For example, it would look for shapes that resembled an eye, a mouth, or a nose, for instance.

And if the order was right, the algorithm concluded that it must be looking at a face. So, to develop an automatic image recognition system, programmers had to describe thousands of images from all angles in machine language and that turned out to be easier said than done. In the traditional approach of classical AI, the machine was fed with knowledge. But it turns out that Deep Learning works much better because instead of telling it how to process the information, the work is left to the computer. Deep Learning has its roots in cybernetics, an area of research where computer scientists often look to neuroscience for inspiration.

Using this method, programmers no longer describe to the machine what a face looks like; instead, they ask it to find out on its own. The system resembles an extensive network of connections that mimic the neurons in our brains. This artificial neural network allows for a variety of adjustments to strengthen or weaken the signals between the links, culminating in an output signal that provides the answer to the question such as, is there a face in the picture. One of the advantages of deep learning systems is that they can work directly with the raw material it gets from the sensors. If it's a camera, the grayscale or intensity of all the colors is measured for each pixel.

If there are 1,000 per 1,000 pixels, for example, the computer processes a million numbers. Each pixel first sends a signal to the network that varies in intensity depending on its brightness. In so-called supervised learning, the machine tests billions of possible settings until it finally gets the answer that the programmers are looking for, and an output signal like "face detected" lights up.

Once this combination is found, the settings are locked and the learning process is finished. The parameters of the model are adjusted so that it eventually gives the known and expected answer. They imitate the examples given to them by humans. It's very fascinating. The mathematicians have been obsessed with trying to figure out why it works, and nobody's really sure, to be honest, why exactly deep learning succeeded. What makes these neural networks so special is that they can recognize the generic shape of a face within a larger image.

The machine is trained by showing it thousands and thousands of images with faces in them - until the perfect setting is found. And from then on, the system identifies all pixel configurations that correspond to a face, while filtering out all other objects. Such systems can now be found in cameras that automatically focus on faces. In video surveillance rooms, readers for postal codes or license plates, In apps for identifying flowers or dog breeds and in the body scanners at airports. Researchers at the University of Michigan wanted to find out how capable these systems are when the objects' appearance is slightly altered. While the system detects a vulture here a small rotation And it sees an orangutan.

This piqued the scientists’ interest in knowing whether self-driving cars might be thrown off by road signs that had been tampered with. They placed stickers on a stop sign, and sure enough, this confused the vehicles' neural networks, causing them to mistake it for a speed limit sign instead. These kinds of errors may explain why machine image processing systems still do not work in critical applications.

In clinics where automatic readers are being tested, decisions are not made by the AI. Those are left to radiologists and doctors, who continuously monitor and train the systems. These are very fragile systems that are only useful when applied to images that are very close to the training data. So, if you have patients with one population, or you use data from one equipment to train the AI systems, they don't necessarily work when you bring them to a different setting. And humans are different.

Humans have a very nice systems level way of thinking about things. They can think about things that are not in the database. They can think about how the model is working and whether or not they want to trust it in a way that these AI systems by themselves can't do. We tend to anthropomorphize these systems and think that a deep learning system can provide a description of what is happening in an image. We think the model understands what is in the image.

But, the way the model associates an image with text is something completely different than when we humans look at an image and describe it with words. These systems ‘general knowledge of the world is incomplete by definition. They lack the bodily experience, the visual experience, the connection between the words and what they refer to in the real world. Until we succeed in including this side, such systems will remain deficient. Humans ascribe meaning to things through experiencing them: Like feeling the force of the jaws as they bite down; the incisors piercing the smooth skin; and the juice squirting out and running down the throat all of which plays a part in gradually defining what an apple is.

For a computer system, on the other hand, it's just a sequence of pixels linked to textual information. Apple! Yet despite their rudimentary perceptual systems, advances in automatic image recognition have revived the dream that machines will one day develop emotions and be able to help us in perceiving the most secret feelings of our fellow human beings. Will artificial intelligence finally be able to give us an objective answer about Mona Lisa’s feelings for Da Vinci or what went through Monika’s head, when her gaze captivated the minds of critics or Jon Snow’s, in his dying moment? How would such an automatic emotion detector actually work? The first step would be to create a list of emotions from the convoluted, infinite variety of our states of mind. In this sense, the research of American ethnologist Paul Ekman has been particularly helpful for programmers.

After a field trip to Papua New Guinea, the scientist came to the conclusion that humanity shares six universal emotions, which inevitably can be read on our faces: Joy and sadness - disgust and anger - surprise and fear. Ekman's theories have even inspired a television series in which a master detective identifies perpetrators based on their micro expressions. The classification is disputed among scientists, but nevertheless serves as the basis for all emotion recognition computer systems precisely because of its simplicity. The six universal emotions serve as the basis.

The next step is then to have thousands of faces assigned to these six categories with humans doing the selection. That’s how the training data for the machines is created. Machine learning continues until the computer system produces roughly the same results as human selectors: joy, sadness, delight, surprise, fear, anger, rage, disgust, disgust, disgust! Once the best setting is found, the systems become toolkits that programmers all around the world use as universal emotion detectors.

The problem with emotion detection is how it’s being used One application is within management of human resources. Over the past few years, service providers have been using chatbots to evaluate job applicants. If you're an employer, you might have people being interviewed by a computer and then you could have the perceived emotion of the system being conveyed back to the potential employer. So those types of things make me a little bit nervous.

Say I was a job applicant, right? And then I have this emotion recognition trained on my face and based on the tone of my voice, based on the way my mouth moves, like one eyebrow is a little higher or whatever, they make claims about whether I'll be a good worker, whether I have a steady personality you know, things that are making really deep claims about like my interior life. This is pseudoscience, right? This does not work Are you in love with me? There must be a bug somewhere. You are bugging me, Gordon! I can easily imagine that the advertising industry could use such a tool to better influence us.

It's also being used in classrooms to detect attentiveness in students. Some have even floated the idea of using it to build lie detectors. That means you could check suspects with such systems to determine whether a person is lying or not. And that could ultimately determine whether that person remains free or not.

Emotion recognition systems often combine micro expressions and tone of voice for their analysis. But when it comes to Mona Lisa, Monika and Jon Snow, the emotion detectors from Google, Amazon and Microsoft all reached the same clear conclusion: they felt absolutely nothing. And since artificial intelligence has no taste buds it has no idea how delicious pastry can evoke memories of a deceased aunt. It has never felt what it’s like to have adrenaline rushing through your body, or tears welling up in your eyes, or your nose runs.

It’s not afraid of anything, doesn’t get goosebumps, knows neither pain nor pleasure, has no own opinion on abstract art, and carries no repressed traumas. In other words, it has nothing of its own to express. Sophia, a superstar among humanoid robots, nevertheless seems to prove that machines can acquire the ability to speak. But Sophia does not know what she is talking about from her own experience.

Her conversations spring from her programming and input from her conversational partners. Like here at the UN, participating in an event to promote technological development. The relationship between Sophia and artificial intelligence is something like the relationship between a magic trick and physics research. I'm not at all impressed by Sophia, I think it's all a bit farcical. I’m more dismayed by how many people are willing to ascribe something like intelligence, emotions and the like, to this apparatus, and the extent to which they’re willing to play the game of someone whom I actually consider a charlatan.

Humanoid robots can be seen as the embodiment of a dream held by an industry obsessed by pairing human logic with computer logic. But the real strength of artificial neural networks might not be its knack for mimicking us, but rather its capability to assimilate vast databases and to sort through these, analyze, and derive correlations that can help us better understand complicated phenomena such as the ground states of elementary particles in a quantum magnetic field, or the factors of air pressure and humidity in the formation of cumulonimbus clouds. There’s no denying that there are many tasks where machines just outperform humans. No human can take a very large database into their head and make accurate predictions from it. It's just simply not possible.

AI is becoming a scientific tool for researchers in many fields including healthcare. It is precisely this ability of mainframe computers, to detect unknown risk factors for diseases, or the effect of a new drug, by scanning huge patient databases, that explains the tech giants’ battle for data from the extremely lucrative healthcare sector. Google used to be a search engine, but now it has a healthcare branch and the same goes for Apple.

And no one is asking why! It has nothing to do with their original business model. This is all connected to Big Data. It all started when two young computer geeks invented the search engine that most of us keep using to quench our thirst for information.

The invention of personalized ads turned Google's founders into billionaires, while their algorithms tuned user data into tell-all biographies. These algorithms are as secret as the recipe for Coca-Cola. All that is known is that they are based on two cleverly applied processes: The first of which is profiling, where all known information about a person is brought together to create their user profile: Internet searches, online purchases, streamed videos, messages sent, and places visited provide increasingly precise clues about what products might interest a person. The second process is mapping, which involves grouping users with the same preferences together. From the moment a user begins to browse the web, jumping from one page to the next their preference for everything from music to politics, and shoes is carefully mapped out. These connections grow clearer as more users make similar choices like a forest path becoming more visible over time.

Anonymous scanning of millions of users creates virtual communities and sociological maps from which digital platforms deduce what other products shoppers might like. You collect a certain amount of information about a person and then try to predict how he or she will act in the future based on how they have acted in the past. Always it is using data from the past to understand the present and future based on a model created from the past. To understand the effect of these algorithms, researchers at Boston University invented ads and distributed them through Facebook's own advertising platform. It turned out that 80% of those who were shown an ad for a country music album were white, while 85% of those targeted for a similar ad for hip-hop were black.

Fictional job ads got similar results: 90% of those targeted to embark on a new career as a lumberjack were men, but 85% women - when it came to a supermarket job. Just as 75% of those who saw an ad looking for new taxi drivers were African American. Social psychological analysis tools for targeted advertising are aimed at probing our stereotypical behaviors to better exploit and thus reinforce them. But these same tools also form the basis for systems that analyze our behavior to guide our collective and personal decisions such as dating apps that suggest partners that best match our profile. Software that helps banks and insurance companies to identify the type of people who might not pay back their loans or programs that pinpoint dangerous areas and calculate the probability of a crime taking place, on behalf of the police.

Whether advertising sports shoes or predicting crimes, these systems make general statements based on the data they have been trained with. There is always a bias with these systems. And one reason for that would be biased training data. Take, for instance, a system that looks at surveillance videos and tries to filter out suspicious people.

To develop such a system, you would first have to ask “Mr. Smith” to watch surveillance videos and decide which people look suspicious. Of course, the outcome is then based on Mr. Smith's judgment,

which may be biased. And as long as it is Mr. Smith who is talking, you are aware that he is expressing his opinion. But as soon as you use his conclusions to train a deep learning system, the bias is no longer obvious.

The social, psychological, and moral context will remain incomprehensible to computers for years to come. But if these systems lack judgment, on what do they base their decisions? On the basis of statistical measures. And those - as studies have proven lead to racial and gender discrimination. In a dozen US states, software informs judges during trials about the risk of a defendant reoffending. This machine learning system has been trained with police offender databases. It analyzes 137 criteria according to a secret formula and delivers its verdict to the judge in the form of a short summary.

Investigative journalists compared the results of 7,000 people and the software's predictions with what actually happened in subsequent years: Only 20% of predictions for serious crimes proved accurate. Most importantly, the journalists found that predictions of recidivism for black people were much higher than what turned out to be the case, while they were too low for the white demographic. So, if you have a racist criminal justice system, as we do in the US and, you know, in much of the world, you will have anti-black bias built into the data. Ultimately, the machine only provides a potentially fallible and imperfect assessment of the person. So, in a way, the bias is being whitewashed a bit like with money laundering.

So how do we handle this modern dilemma? Should those who are given a choice, rather trust machines or humans? Just as machines might lack empathy, we humans are often not the best at math. And we can be: emotional, immature, sleepy, lazy, rebellious fun-loving, overworked. Or even completely delusional. All these technologies fit perfectly with our - quote-unquote “fundamental human laziness" Because in today's world, such systems offer us a convenience by taking over part of our daily chores.

Our biggest challenge right now, is to take control of our individual and collective destiny. Yet systems are doing just the opposite in many areas of society. Originally intended as a memory aid, AI is now making recommendations and even automatic decisions. Meanwhile, machine learning systems, with their billions and billions of possible settings, are so complex that even the programmers no longer understand the criteria on which the machine is basing its judgment.

A term has been coined for this phenomenon: Blackbox A black box machine learning model. It's a predictive model that is either so complicated that a human cannot understand it, or it's proprietary, which means we have no way of getting inside and understanding what those calculations are. Especially with deep learning systems, it is not clear how decisions are made, since only the results are visible. There is a movement where people are saying: Well, we can still use black boxes, we just need to explain what they're doing. And so, I've been trying to kind of beat that back down and say: No, no, no guys, we don't need black boxes! You can't just launch a black box. That's a high stakes decision.

You actually really need to understand how these predictive models are working in order to make better decisions. Behind user-friendly interfaces there is often a closed decision support system. Apart from that, it’s well known that companies like to avoid responsibility, by saying something like: We are dealing with a very complex system here and It wasn’t us. It was the algorithm. That kind of reasoning or excuse is of course completely unacceptable.

It's a very good way to sort of evade a responsibility and make difficult decisions that you may not want attributed to you: It was the machine! A fundamental value that defines us is freedom of thought, which can be traced back to the Enlightenment. But now trade-offs are being made, as we are delegating our decision-making. The underlying goal is to prevent any errors - but to do this, we are handing the control over to systems. The self-driving car has long been the poster child of artificial intelligence, epitomizing this dream of global automation serving humanity by making the decisions for us, all the while keeping us safe and relieving us both of pressure and potential road rage. But despite investments of almost 100 billion euros, no self-driving car has yet been allowed into traffic outside of the test tracks without a human driver ready to grab the wheel at any given moment. It is very easy to use deep learning to make an unreliable prototype for something very complex - like driving.

But it is very hard to develop this prototype further so that it becomes reliable and powerful enough to be practical in traffic, for example. For economic, political or even technical reasons, it is the system itself that requires products relying on artificial intelligence to be brought to market while still not functional, or only partially functional. In Silicon Valley, they say: fake it till you make it.

You have to pretend it works until it finally does. But that time is constantly being pushed back because artificial intelligence never reaches a level of performance that makes humans dispensable. The predictable failure of autonomous driving has spawned a new profession: human assistant to machines in need. This company, for example, trains employees to take control of not-quite-autonomous vehicles. In just ten years, an entire industry has sprung up around artificial intelligence assistance.

Hundreds of thousands of workers around the world prepare data for training of machines, checking and correcting their answers or simply replacing them when needed. This new form of human labor behind so-called "artificial intelligent systems" was invented by the billionaire founder of tech giant Amazon. When Jeff Bezos announced the launch of Amazon Mechanical Turk in 2005, he made no secret of the fact that it was a project for "artificial artificial intelligence," that is pretend artificial intelligence. In this case humans are doing the necessary manual labor, so to speak, to make these algorithms work.

Amazon Mechanical Turk is a crowdsourcing platform that addresses this apparent paradox: the growing prevalence of automated decision-making systems and the inability of artificial intelligence to be truly autonomous. The platform is named after an 18th century automaton, the so-called "Mechanical Turk" with a human hiding in its base. It literally refers to hidden labor, right? We're talking So yeah, it's like they do say the quiet part loud often. We’re looking at a paradoxical situation: on the one hand, people are being asked to do what robots or automated processes are not capable of doing, while on the other hand, workers are tasked with activities that give them little wiggle room and much less autonomy than before. Science fiction has stoked up our fear that computers could become so intelligent that they succeed in dominating us like Stanly Kubrick’s dangerous HAL. This machine is too important for me to allow you to jeopardize it.

But currently a very different sentiment is taking hold: It’s not so much that the machines have become so intelligent that they are dominating humans. But rather that we are gradually submitting to the standardized logic of the machines. In call centers like this one, employees have to adapt to an algorithm, and are increasingly being monitored by artificial intelligence. Software identifies keywords in the conversation to ensure instructions are being followed.

Emotions of the customer and the employee are analyzed in real time. If anger is detected, the system directly prompts the employee to be more empathetic. At the end of the conversation, employee performance is rated and if it falls below a certain score they are immediately fired.

We are witnessing the dawn of an age in which multitudes of people are forced to adapt to the dynamics of interpretive systems AI systems designed to maximize optimization and productivity, allowing no room for negotiation. In reality the Amazon warehouses are dystopian nightmares where employees have to try to keep up with a partially automated system. In the vast halls of distribution centers, the computing power of artificial intelligence ultimately seems to offer little help when it comes to understanding the complexities of the real world and making the best decisions.

Here, the work done by humans is subject to the commands of machines that only calculate the optimization of the flow of goods. They are the ones setting the pace. People are reduced to being robots of flesh and blood.

In my opinion, automation is a trick to control enormous masses of people. My biggest concern is not computer bugs, but people who seek power, who want to control others, and who increasingly have access to very powerful technologies. There's always hope, you know, a hopeful outcome is inevitable, and I think it's going to take work, right? Like hope is an invitation not a guarantee.

2023-05-27

Show video