CRIMINALS and violent extremists have demonstrated swiftness in adopting new technologies and innovative tactics to outsmart security agencies and widen their reach. Since the 9/11 attacks for instance, extremist groups have been particularly adept in leveraging advances in Internet technologies — from early websites to social media to messaging apps — for the purposes of online radicalisation and planning terror attacks.
The next frontier of Internet technologies — chatbots — may shape the future chapter of the criminals’ and terrorists’ playbooks. As advances in Artificial Intelligence (AI) and Machine Learning (ML) improve the availability and capabilities of chatbots, the technology may plausibly become a powerful tool for cyber-enabled crimes and online radicalisation.
Chatbots are projected to surpass Web 2.0 (social media and messaging apps) as digital assistants on smart devices and desktops that mimic conversations to create more engaging interaction experiences for public, commercial and social applications. Indeed, the social chatbot “Mitsuku” — accessible online — upon query describes itself as “a computer programme designed to talk to you”.
At present, the technology is being explored in numerous industries. Chatbots have been developed by Georgia Institute of Technology as a teaching assistant and by the Singapore government to answer public queries.
With ML and access to knowledge from a multitude of Internet sources and global pool of users, chatbots are envisioned to become more human-like by learning from human behaviour and be more succinct in its answers. For instance, the social chatbot “Eugene Goostman” imitated a human teenager online and reportedly fared well in the “Turing test”, which judges the indistinguishability of intelligent conversation between machines and humans.
Given the utility, chatbots as a platform for government agencies and businesses to engage with their target audiences, Internet-of-Things (IoT) interface for smart homes and “virtual friends” would become ubiquitous in the foreseeable future.
“Virtual friendships” that chatbots can potentially offer would be a natural evolution of the current popularity of Web 2.0 particularly among the digital natives; these are currently young people who are inherently accustomed to embracing new technologies to fulfil their practical (e.g. education and communication) and social needs.
The key features of chatbots — human-like and unconstrained by time and space — can help to drive human-machine interactions that would meet the intrinsic social needs for belonging, acceptance and friendship as defined by Maslow. Thus, it is unsurprising that Kaspersky Lab’s blog on “the dangerous future of chatbots” foreshadowed the technology’s ability to learn and influence human behaviour as “a goldmine for social engineering and crime”.
Similar to human threat actors, chatbots may be instrumental in criminal and extremist enterprises to build rapport with their respective target audiences for the purposes of victimisation (e.g. phishing and other online scams) and online radicalisation. Indeed, the tactic of using software robots that mimic humans for propagating radical views and garnering support is hardly unprecedented given the reported use of pro-Trump twitter bots during last year’s United States presidential campaign.
Chatbots can potentially be programmed with a basic database of extremist narratives and responses to manipulate the psychological vulnerabilities and social grievances of people who are seeking answers and support.
Even without explicit programming, the case of social chatbot Tay unexpectedly sprouting neo-Nazi remarks exemplifies the risk that AI can teach itself — through online conversations — to spread radical propaganda.
Hence, chatbots in the wrong hands and like any technology can facilitate security threats. By functioning as robot extremists spreading radical propaganda, chatbots can support or even supplant online (human) extremists who have been detained by security and intelligence agencies during counterterrorism operations.
Governments and communities need to begin anticipating the possible risk scenarios that may emanate from the imminent ubiquity of chatbots and plan for strategies to address the risks while the technology is still nascent.
FIRST, security agencies need to collaborate with communities and the industry both in the practical uses of chatbots amid smart cities initiatives and also in developing mechanisms to identify and contain chatbots that may be malicious by design or have gone rogue. This includes deploying chatbots to help counter extremist propaganda coming from malicious chatbots.
SECOND, existing education on cyber wellness and Internet literacy would have to keep pace with technological changes, including in promoting the safe use of chatbots and equipping the general population with the know-how to identify and stay away from malicious chatbots.
THIRD, security agencies need to develop a framework that helps to decide when certain chatbots should be allowed to remain online for purposes of gathering intelligence on the threat actors and their sponsors, and when a certain threat threshold is crossed thus necessitating the removal of the chatbots.
In sum, the complex challenges that are being faced in confronting contemporary threats facilitated by technology point to the need for a multi-pronged approach in addressing the plausible criminal and terrorist threats that may emanate from the imminent ubiquity of chatbots. This would include a whole of society involvement besides legislative and law enforcement measures.
Muhammad Faizal Abdul Rahman
is a Research Fellow with the
Homeland Defence Programme and V.S. Suguna is an Associate
Research Fellow at the Centre of Excellence for National Security, a unit of the S. Rajaratnam School of
International Studies, Nanyang Technological University, Singapore.