“Study the past if you would define the future” ~Confucius
Somewhere during the Pleistocene Era, an Australopithecus picks up a humerus and discovers it can use it for a decisive advantage over its rivals in conquering a waterhole. It then proceeds on celebrating its triumph, raising its newfound weapon to thrust upwards into space where it transforms into a space station. This famous scene from the film “2001: A Space Odyssey” by Stanley Kubrick magisterially captures how the story of humanity is intimately linked to technical progress, and with warfare as its primary driver.
Today, while the investment in artificial intelligence (AI) is in full swing across most industrial sectors, questions are being raised as to how this seemingly new technology will impact the global economy and society, and how it will influence warfare and our use of weapons. The recent and spectacular advances in the field of AI triggered much excitation, to the point where non-AI experts occupy the forefront of the debate, gaining traction based on their entrepreneurial success, and scientific prowess or their futuristic views. This is how we may be doomed by almighty killer robots, AI-moralizing instances, or even a speculative singular event. All the while, people involved in developing AI, and who are quite busy doing so, are seldom heard. This results in a discussion imbued in a rather phantasmatic, if not a religious tone that steers away from science and actual knowledge; hence, of informed prediction or at least, informed strategy. While grounded and useful positions can be found in military literature, the consensus seems stuck, at least in France, in a “wait and see” mode, if not disappointingly shallow. This is partly due to the relative precociousness of AI development, and to the intellectual ‘fog of war.’
We are at the stage where the monkey is picking up the bone before realizing it could bash skulls in with it. While the deployment of AI within the military is unavoidable, we question what to do with it. This question will not be fully answered in this essay, but we will try and provide recommendations, or at least insights, as to how to harness AI’s power, so that others can pick up the trail and push our collective light to solid grounds. We propose to do so by looking at the past to understand AI’s origins, by taking the measure of its current status, and provide some landmarks of what can be expected in a straightforward and comprehensive manner; this will lead to a prediction as to AI’s impact on cyberspace, which we believe is premiere in its magnitude.
What is Artificial Intelligence?
There are several definitions for AI, but they can all be traced back to a common one: “the simulation of (human) intelligent processes by machines.” This definition is unsatisfying because it relies on the concept of intelligence, which itself is very loose. Stacking cubes one atop the other is a mark of a baby’s intelligence, but how does it relate to the measurement of an adult driving a car? What is common between the intelligence displayed by an expert analyst of satellite images and the intelligence a military leader designing a military strategy? Do we have to leave out the intelligence displayed by a group of ants collectively, and optimally, forming a bridge to cross a river? Relying on the concept of ‘intelligence’ is more misleading than helping.
What we do know is that AI is a technology that is the science of a given technic. We thus propose to consider AI as the “science of computing,” the one that answers the question of “how to process information.” The intelligence of the expert image analyst hence becomes the process of transforming a visual scene of information, as described by a matrix of grayscale values, into the information relative to tanks and troop locations, changes in a building setting, etc. Armed with this definition, we will try and put an operational light onto what is AI. We will dismiss the question of “narrow” AI versus “broad” AI, as every AI so far has been narrow, with broad examples simply imaginary.
The Age of Rules
The birth of AI can be traced to Alan Turing’s paper “Computing Machinery and Intelligence” (1950), later to be baptized at the Dartmouth Workshop (Dartmouth Summer Research Project on Artificial Intelligence in 1956). In this seminal paper, Turing addressed the question of whether a machine thinks, through the introduction of the concept of the “Imitation Game,” in which a machine could be considered as thinking if it fooled a human into believing it was conversing with another individual. The digital computer would do so by “carry[ing] out any operations which could be done by a human computer,” the latter described as using a defined book of rules. The core of Machine Intelligence is found in formalization of thought as logical rules, of “IF-THEN-ELSE” that are the building blocks of all programming languages up to this day. Modern computers are hence artifacts of the first wave of operational AI that defines intelligence as following a given set of rules to solve a problem. The hallmark of this approach is exemplified in Deep Blue’s victory over Gary Kasparov, arguably the best chess player of all-time. Not that the machine was particularly smart, it was programmed the rules and objectives of the game and had enough computing power and memory to bulldoze its way to victory.
The key to understanding AI is that within the current digital world, a given problem is solved through automation using logical rules that the engineer has applied to the problem. In the field of cybersecurity, for example, a SIEM (“Security Information and Event Management”) detects threats based on rules that experts implement based on their analysis of previous attacks; the experts will have to refine the rules as false alarms are raised when a pattern of events overlapping with the definition appears, and as their analysis of the attack deepens. While the rule-based approach has a positive value in helping analysts in detecting threats, but the example highlights its main drawback; it is strictly limited to the subjective understanding of their designer.
Under the “Rule-based” paradigm, which can be described as one of “Delimited Mastery,” digital models of intelligence explain the known world but fail to extrapolate outside of its boundaries. Additionally, these models are generic and can only adapt to the specificities and changes of their environment through the direct intervention and understanding of a human operator. In a sense, one can say that the intelligence displayed by the machine is strictly inherited from its designer’s knowledge.
The Age of Learning
The recent AI developments explain why it is at the forefront of technological discussions. At the 2012 ImageNet Challenge, Geoffrey Hinton’s team destroyed the competition in the field of Image Analysis thanks to an original information processing architecture, later called “Deep Learning.” The multi-layered neuronal network architecture was on the AI shelves, minus one or two original and decisive mechanisms, but it had finally been able to challenge competitively because of two “logistical” reasons:
- The outbreak of cheap but powerful processing units specialized in parallel processing (GPUs), which could address the substantial needs of the parallel architecture;
- The profusion of data made available by the interconnection of machines and humans, thanks to the democratization of smartphones and social media.
The success at ImageNet Challenge triggered massive investments in the field of AI, as major technology players such as Microsoft and Apple entered the once purely academic competition as early as 2013. The investment led to the Go Game in 2015, when a game traditionally thought of as intractable would be won by a machine against a human master, in spectacular fashion. The event is proof that machines have become smarter and is the main reason people are so excited about AI. The machine discovered and introduced moves that left the experts and commentators baffled. Humans did not explicitly implement the behavior; its designers’ role, not to be underestimated, set up the right conditions for the machine to learn and perform in the game.
A new paradigm is established in information processing that relies on three pillars to be operational, which is define as the “ACD” triangle:
- Algorithms and people apt at designing them;
- Non-trivial Computing power;
- Massive amount of Data.
Humans no longer must specify the model by which to address a given task. Instead, their task is to design an architecture and provide a sufficient amount of relevant data for the machine to learn its own model, possibly discovering relationships in the data that would have escaped the expert’s intelligence:
- Because the number of information channels to analyze simultaneously is too large (spatial dimension);
- Because the signal is too weak to be perceived based on a single observation (temporal dimension);
- Because information channels will be taken into consideration that could have been discarded by an engineering bias (cognitive dimension).
As a result, the machine will be as useful as the data accurately covers its complexity; as such, it can potentially be below an expert’s knowledge, and in practice, the machine performs on par. The derived model will also address the unknown world during its learning process. The AI model associates configurations within the data with classes (cats and dogs in the case of Image Analysis for example). Once the learning is completed, new input will be presented that is entirely different from what the system was taught (e.g., Siamese cats, or old dogs). It will nonetheless produce estimates, typically under the form of probabilities across its class vector; the better the estimate, the closer the machine will reach the “idea” of a given class (e.g., “Cat-ness”). Said otherwise, Machine Learning opens the way for machines to generalize, reaching beyond what it learned to catch what is similar, provided the ACD triangle is respected.
Going back to the cybersecurity example of Threat Detection, and applying it within the Learning paradigm, the apparent strategy would be to try and collect all the data corresponding to all the attacks and have a machine learn a general mode of attack. While theoretically possible, the data collection would prove difficult, if not impossible, as the attack topology is extensive and changing. An alternate strategy would be instead of collecting the data relative to the life of the Information System (IS), and have a machine learn a model of the IS’s nominal behavior that would point out anomalies among which attacks, because they, by definition, do not correspond to the expected behavior of the system. And there you have EBA (Entity Behavior Analysis).
Under the Learning paradigm, which can be described as one of “Educated Guess,” AI designers must specify an architecture and provide data, which leaves room for their biases to alter the system’s quality. The machines do not inherit their intelligence as they build their model of the world. As such, they become able to generalize beyond what they learned, which was not the case in the previous setting, or by accident. Also, models become inherently specific to their environment. The same architecture will not develop the same model in two different infosystems; one model produced in environment A will not be as sharp as the one in environment B.
What this could mean for Cyberspace?
The machines are smarter within the Learning paradigm than within the Rules’ because the Age of Rules peaked when it captured the Learning process itself and faded as it deployed its successor. To consider that AI would surpass, or even match, human intelligence is, at best, speculative. This would mean that we would be able to replicate the brain, which is the seat of human intelligence, and neurosciences tell us that we are, at best, far from it. Would it be accurate to surmise that the debate is merely pointless; nothing humans could say or do would matter if machines become more intelligent than us.
It is the author’s opinion that human and machine intelligence is non-comparable because they are different. They are similar in that one draws from the other. What Kubrick’s scene suggests is that humankind appeared when monkeys started coevolving with their tools and weapons; humans shape tools and tools change the way we interact with the world. AI is a technology that is the fulcrum of the information processing tools and weapons that we will choose to design and build. How will AI impact humans? It can synthesize and make its own sense of the overwhelming amount of data humans produce through digital systems. As such, AI has the potential to act as a therapy to the Big Data disease. Said otherwise, humans can no longer need to cope cognitively with the amount, rate, and heterogeneity of information they must deal with in a hyperconnected world; therefore, our vision is one of AI acting as cognitive support to humans.
Humans build weapons and weapons change the battlefield. AI will undoubtedly improve the speed and accuracy of various weapon systems, such is the quantitative value of automation. It would also have a qualitative impact of making weapons smarter; being more specific is difficult because the results will be as diverse as the military devices it is applied to. But think of a rifle with a scope able to indicate whether your target is right- or left-handed so that you can aim at his actual weapon-directing arm when he is not holding it. The key here is thinking specific (this opponent’s main arm) where generic was the norm (any opponent’s right arm).
Now, one aspect we voluntarily left out so far is the machine’s operational domain. It is remarkable that within the first paradigm, machines were operating only inside the digital world. Rule-based algorithms in Machine Vision could hardly analyze and extract much information about natural scenes. Deep Learning changed that construct. Machines were made to talk to one another through TCP/IP, leading to the Internet and global interconnection, effectively increasing the reach of the digital world in the process, yet discussing with a machine in a natural language was impossible. Thanks to recent advances in AI, Natural Language Processing has made tremendous strides that hint at conversations naturally flowing with the machine sometime soon. What we are witnessing is a change to the AI paradigm with the extension of the machine’s realm into the “natural” world.
In the cyber field, humans design machines to target other machines to indirectly impact humans and institutions. What happens when data describing the personality and the psychological levers of people is made available for a machine to process and learn from? What are emails and social networks messages, if not the expression of the human psyche? What if an attacker programs AI to detect a weakened employee in a major corporation, and then tailors a campaign of artificial social exchanges to ‘take control’ of the employee? Because of the massive interconnection between humans and machines, and because machines become smarter with increased data, AI-enhanced cyberattacks have the potential to directly strike at humans, and control them from a distance, through machines. This will have consequences on the battlefield if the soldier is heavily dependent on technology.
The future of AI will consist of cyber cognitive threats. This new environment can also be one of cognitive support. While we are picking up the bone, we should be aware of its nature, its origin, to evaluate its aim; this humerus has everything to do with information processing. In the end, “wait and see” can mean having to submit to it, instead of taking control and shaping it according to our will.
Bio
Rudy GUYONNEAU is a Senior Consultant in Artificial Intelligence for Cybersecurity at Sopra Steria SA. His doctoral thesis with Spikenet Technology at the Brain and Cognition Lab in Toulouse led him to study visual information processing in the primate’s brain, and to apply its keys to the fields of Machine Vision and Neuromorphic Engineering. After a postdoc at Georgetown University in Brain-Computer Interfaces, he led the R&D effort at Spikenet, applying spiking neural networks and related technologies to the industry, including video protection. He joined Sopra Steria in 2016 to aid the development of AI within cybersecurity, and to assist Airbus Commercial Aircraft in its Innovation Strategy. He holds a Ph.D. in Computational Neurosciences from Toulouse III University, and is the Lead Cyber Data Scientist for Airbus CA.