Contributed Commentary by Pieter Ballon, Director of SMIT at VUB
Science fiction films featuring robots or intelligent machines in the leading roles (such as Blade Runner, Real Humans, Westworld, etc.) have caused us to look at a future with AI with some trepidation. But it won’t happen overnight. We will have time to adjust ourselves to the idea and to control AI systems where necessary so that it becomes a gradual evolution, not a sudden revolution. But it is definitely an evolution that is already underway.
Harvard Prof. Michael Porter, sets out four stages that mark the way toward smart objects and systems. Stage one is “Monitoring” by using sensors: a smart product will be aware of its own situation and the world around it. An example of this is the Medtronic glucose meter, which uses a subcutaneous sensor to measure a patient’s blood-sugar level, alerting the patient 30 minutes before that level reaches an alarming status.
Stage two is “Control”. Thanks to its in-built algorithms, the product will then carry out an action based on the readings or measurements it has taken. For example, if a smart camera detects a car with a specific number plate, the gate will open.
Systems then evolve towards the stage of “Optimization”. Using all the data that the system collects while it is operating, in-built algorithms can carry out analyses to determine the best way of working. It’s as though the system “learns” to work more efficiently. An example of this are wind turbines that are able to adjust the position of their vanes each time the wind changes direction so that they can capture a maximum amount of wind energy and also disturb the flow of the wind to any neighboring wind turbines as little as possible.
Finally, smart systems evolve toward “Autonomy”. When a product is capable of monitoring itself or carrying out an action—making that action as optimal as possible—it can work autonomously. For instance, the iRobot vacuum cleaner robot is capable of cleaning all sorts of surfaces in the home, as well as detecting dirt, finding its way round furniture and avoiding tumbling down stairs. It also stores details of the layout of a room in its memory for the next time and makes its own way back to its recharging station, where it announces its safe arrival with a triumphant sound signal!
Smart systems can also be connected with each other so that they can carry out actions in tandem, learn from each other—and so on. Driverless cars and road infrastructure can work together so that if there is an accident somewhere, cars further away from the incident can be notified and the appropriate action taken.
As we can see from these examples, we will gradually evolve toward systems that are capable of learning and making decisions by themselves. And equally gradually, we humans will hand over the monitoring, control, and optimization—partly or in full—to machines.
People and AI Systems Will Become Workmates
Human-like AI, human-centric AI, human-in-the-loop AI—these are all terms to indicate that human beings are still very much central to the story. Robots and machines need to be made in such a way that people can understand them, are able to communicate with them and can work efficiently with them. That way, machines can carry out tasks on behalf of and for the benefit of humans.
This will happen faster in some sectors than in others and there are various reasons for that. In the mining industry, for instance, Joy Global’s Longwall Mining System is used to dig underground virtually automatically, without any human input. Staff sitting in the control room above the ground keep a close eye on everything going on and only send engineers below ground if it becomes necessary.
Another good example is the ‘cobot’, or collaborative robot, developed to assist Audi production line workers in assembling cars. Whereas previously these types of machines used to be placed in safety cages, the cobot is able to carry out certain actions safely close to and with the help of its human workmates. This means that tasks such as applying adhesive can be carried out much more precisely, consistently and always in the same way. Meanwhile, the cobot’s human workmate is able to control and direct it using hand gestures.
Matter of Interpretation
There are still many challenges to overcome in the area of communication between robots and humans. For example, will a robot ever be capable of identifying our intentions? Can a robot detect if we say something in a fearful or more self-assured way? Or when we carry out an action, what does this say about our actual intentions? For instance, it is no easy task to get a driverless car to recognize whether a pedestrian intends to cross the road, or is simply standing at the side of the road. Typically, as a pedestrian, we will try to make eye contact with the driver to indicate that we would like to cross. But this type of “unwritten rule” in human-to-human communication is not easy to transfer to AI systems.
For cobots and humans working together, perhaps the human demonstrates how something is done and the cobot learns from it so that it can then perform a particular action perfectly. In a single, repetitive process it may be that the human worker will only have to show the robot how to do something a few times and the robot will then take it from there. But in more complex situations, the cobot may always need a human workmate on hand to give instructions and to model how to do things.
For example, it can be a complex business collecting trash in a city and distinguishing what is waste and what isn’t. It’s also hard to know how to react if someone waves to the waste truck driver and then runs up behind with a bag of waste to be picked up. A robot would not know how to respond, whereas a human knows that the friendliest thing to do is wait.
In situations like these, humans and robots may always have to work together, with the robot taking on the heavier work and its human workmate having more time for interaction with other people and responding to unexpected situations.
AI Systems Need to be Tested Regularly
As we have already said, the human needs to understand how a robot arrives at a certain conclusion or action—and must always be able to make adjustments where necessary. Recent examples of problems with artificial systems have demonstrated exactly that. For example, the chatbot Tay began posting racist messages on Twitter after certain other Twitter users left politically incorrect posts. The chatbot had not been given any instructions to recognize these types of statements as being inappropriate.
‘Norman’ also made the news in 2018. Norman is an AI system that displayed psychopathic characteristics when doing a well-known test with Rorschach inkblots. It happened because Norman had previously been shown mainly sensational and violent images from Reddit and he had built up a picture of the world based on those images. MIT researchers wanted to use the experiment to demonstrate the danger of “false data” being used as input for AI systems.
And finally, there is also the example of the COMPAS algorithm that was used by the judicial system in America to make predictions about the recidivism of convicts. What happened? Based on the historical data used as input for the algorithm, it reached the conclusion that blacks were more likely to re-offend than whites.
All this has taught us that we need to look very carefully at the data used as input for AI systems. It also shows us that it is important to understand at all times how an algorithm arrives at a certain conclusion; that we can make adjustments to it. The AI system cannot be a “black box”. New procedures and checks are required to ensure data and algorithmic transparency.
Yet it will probably not be possible to make all data input and algorithms wholly transparent—which means there will be times when we do not know why an AI system may have come to certain conclusions. So it is important to introduce regular tests for AI systems, which include paying a great deal of attention to ethical issues, so that any problems can be identified quickly.
How Can We Trust ‘Them’?
So, how can we ensure that everyone is able to trust the AI system they come across on the work floor, at the doctor’s surgery, or on the road? First and foremost, by allocating some sort of approval certificate, based on a regular audit of the system. One example of this are the elevators that we use every day. We have no idea about how their technology works, but we trust them to be safe and work properly, and we can do this because they are checked regularly by people who know what they are doing.
The way humans and AI systems communicate and their predictability can also help to build a relationship of trust. For instance, take the traffic lights that we use to cross the road as pedestrians: they have a button that enables us to provide our own input while a camera should also be able to detect pedestrians. We receive a signal that our input has been registered; next, we wait patiently because we know that we are using a predictable system that will turn green within a maximum of 2 minutes; and in some cases, the system even tells us how long we will have to wait before crossing.
We need to try and achieve the same level of trust in AI systems that we already have in smart traffic lights.
Will We Still Have Jobs in 2035?
There is a good chance that many routine jobs will be taken over by AI systems and robots. That may range from working on a conveyor belt in a factory, to making certain medical diagnoses, or working as an accountant or in legal jobs.
Even tasks where a bit more creativity is required can be carried out by AI systems. The well-known historian, Yuval Harari, quotes the example of Google’s AlphaZero, which took less than 4 hours to learn how to play chess, after which it was able to beat the best human-trained chess computer not by learning from historical data, but by using machine learning to teach itself to play the game. Now when players in chess tournaments make a move that is strikingly creative and original, the judges may suspect that player is using a chess computer to come up with the moves.
So, both routine jobs and jobs in which new possibilities have to be explored can be carried out by AI systems. But will they ever be as creative as we, humans? That remains to be seen.
Whatever happens, the content of many jobs will be changed by the arrival of AI. We will work with these AI systems and have to keep adjusting to new capabilities. Lifelong learning will be very important, both for low-skilled and highly qualified jobs. New types of job will also be created that we can’t predict at the moment. Typically, jobs where human contact is very important—such as nursing—will still be done by humans, even though robots may be brought in to provide assistance and support.
But the impact goes further than just our job content. We will also find other economic models becoming involved in which we will work less and in which job, income, and consumption will be separated out from one another. Perhaps you won’t have to work (fulltime) to receive a full income—and maybe you won’t need a full income to provide for your basic needs. Just look at the current digital economy of apps and digital services. You can use many services free of charge and sometimes (with or without you realizing it) in exchange for access to your data. With this system, data will become a new tradeable commodity and so we will be able to ensure better that our data is our own property and that we can use it transparently when and where we want.
Pieter Ballon is director of the Studies in Media, Innovation and Technology (SMIT) group at The Vrije Universiteit Brussel (VUB), a Dutch-speaking university located in Brussels, Belgium, that emphasizes engineers and social scientists working together on AI, seen as a technological innovation poised to cause significant economic disruption and social changes. He can be reached at [email protected].