It is predicted that investment in AI could add £630 billion to the UK economy by 2035, increasing the annual growth rate by up to 3.9% [1]. The adoption of AI, however, often represents a significant cultural change for the industry or organisation in question. Ultimately then, AI adoption is unlikely to succeed without buy-in from practitioners. In short; people need to trust AI before it can realise its full potential. This document summarises how certain technical aspects of AI implementations can affect the level of trust that practitioners place in the technology.
A model is described as ‘overfitted’ if it fits well to the training data but fails to generalise to situations that are outside of the training data. Typically, models with more ‘tuning parameters’ are more prone to overfitting; a polynomial with parameters, for example, can be made to exactly fit to data points but will likely extrapolate poorly beyond (or even between) the training data.
Deep Learning neural networks, which have recently been the subject of considerable attention, can have as many as 10^9 parameters [2] and are, as a result, prone to overfitting. When extrapolated beyond training data, overfitted models will often generate predictive results that appear bizarre to humans. A recent paper [3], for example, illustrate examples where a Deep Neural Network has been trained to classify images but fails to generalise when the objects-in-question have been perturbed, an identification task that is clearly second-nature to human beings.
Overfitting can be mitigated by the use of regularisation schemes (which penalise a model for being overly complex) as well as sensible validation metrics that aim to test a model’s ability to generalise outside of the training data (k-fold validation, for example). Models that embody uncertainty quantification and expand their predictive confidence bounds to indicate when they are being asked to make predictions in regions that are not represented by the training data (Gaussian Processes, for example) don’t necessarily solve the problem but they do at least let you know when predictive performance has deteriorated (as well as where more data may need to be collected). In my experience, however, the most obvious solution involves adopting simpler models. In many cases, people turn to Deep Learning because of the hype surrounding the field, and not because it is a good choice of model.
"…deep learning techniques thus far have proven to be data hungry, shallow, brittle, and limited in their ability to generalize”
[Marcus, G. (2020). The next decade in AI: four steps towards robust artificial intelligence. arXiv preprint arXiv:2002.06177.]
In most cases, AI is used to tackle relatively basic and highly-repetitive tasks that would take humans a long time to perform. When AI is used to tackle more sophisticated challenges, the scope of the problem is usually very narrow. AI is not, in its current state, able to demonstrate what is commonly referred to “general intelligence”; a 2021 report by GCHQ into the ethics surrounding AI adoption [4] states that “the AI systems we can currently build are only able to tackle very limited, tightly-defined problems, and then only with human support. At least for the foreseeable future, AI engineers will certainly not be creating any intelligent computers that can truly replace humans.” It is therefore important, from the beginning of the adoption process, to manage the expectations of what AI can really do.
Many AI models are presented as “black boxes”, where the mapping from input to output is hidden to the practitioner (another problem that is common to Deep Learning). Even if these models work well, they are difficult to interrogate and essentially exclude the practitioner from the model development process. In my experience, the application of black-box models to real-world problems usually results in awkward situations whereby practitioners (quite rightly) ask the question “why does the model do that?”, to which the AI modelling team has no convincing response. Moreover, not knowing the factors that led an algorithm to make a particular decision may also have serious implications regarding transparency and accountability [4].
Having a model that can be interrogated (i.e. explainable AI) raises the possibility that practitioners may be able to learn, themselves, about fundamental phenomena that have been revealed by the model, combining the ability of AI to find patterns in high-dimensional datasets with human beings’ ability to generalise. This possibility helps to involve practitioners in the model development process, therefore developing trust in the approach as well as an understanding of its limitations.
It is also worth noting that the use of black-box approaches can lead to vendor lock-in; the 2020 government guidelines of AI procurement [5] features an entire section titled “Avoid Black Box algorithms and vendor lock-in”, which contains the text: “Encourage explainability and interpretability of algorithm and make this one of your design criteria... This will also make it more likely for you to be able to engage with other suppliers to continue to build upon your AI system in the future, limiting the risk of vendor lock-in”.
The need for explainability has led to some calling for a return to so-called “symbolic” approaches to AI
[6] which have been around since at least the early 90s
[7].
"Although machine learning methods are solving very successfully a plethora of tasks, they have in most cases the disadvantage of acting as a black box, not providing any information about what made them arrive at a particular decision”
[Bach, S., Binder, A., Montavon, G., Klauschen, F., Müller, K. R., & Samek, W. (2015). On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PloS one, 10(7).]
Utilising approaches that can leverage the knowledge of domain-experts is key; it is very difficult to build trust in algorithms that only learn from data and are unable to incorporate domain-level expertise. As an example, we consider Gaussian Processes (GPs). GPs are an extremely flexible machine learning approach that automatically facilitate uncertainty quantification and have, as a result, received significant attention from parts of the AI research community in recent years. It can, however, be surprisingly challenging to incorporate basic domain-level knowledge into GP models. It is quite difficult, for example, to implement a GP model that is limited to only produce predictions that are greater than zero; this can be a little embarrassing when it becomes apparent that this advanced AI approach struggles with the concept that, when counting numbers items, it isn’t possible to realise a negative result.
Reinforcement Learning (RL) involves placing an AI “agent” in an environment (which could be real or simulated). The agent then performs actions in this environment and is penalised or rewarded for the resulting outcomes. In this way, the agent learns how best to interact with the environment; a “policy” that maximises reward.
RL is particularly well suited to playing games whereby the agent is a player, the environment is defined by the game and the actions are certain moves associated with the game. Famous RL success stories include the development of AI algorithms that can defeat the Go world champion [8] and, more recently, achieve “grandmaster” level in Starcraft II [9].
Robotic learning – robots that learn to interact with their environment - is an obvious area where RL would appear to have great potential. There are, however, issues associated with transitioning RL approaches from games to real-word scenarios that harm their adoption [10]. Firstly, essentially by definition, the RL agent has to make some mistakes (i.e. poor actions) during the training process. During gameplay, a poor action will simply result in the agent losing the game, however, in robotics, a poor action could result in behaviour that is damaging (to the robot and/or the environment) as well as unsafe.
One possible solution involves training the algorithm to recognise unsafe situations and take corrective actions as a result. Alternatively, a large amount of the algorithm training can be conducted within a simulated environment, although a lot of care must be taken to ensure that the resulting approach is robust against the natural stochasticity of the real world (the issues associated with the differences between simulated and real environments are often referred to as the “reality gap”). The UK’s AI roadmap
[11] advocates the use of “living labs”; controlled physical environment where AI solutions can be trialled.
It is important to note that AI algorithms will typically be applied in non-stationary environments, and must therefore be able to learn online. Without this, the practitioner will likely see a gradual degradation in algorithm performance as the time-since-training is increased. Work by Silver, D. et al. [10], for example, reports examples from robotic learning where learned locomotion policies stopped working only a few weeks after training (as a result of hardware degradation etc.) while GCHQ advocates the use of a life-cycle plan to address potential slow declines in AI algorithm performance [4].
AI algorithms are not immune to unfair biases. Algorithms that are trained on skewed datasets may end up working better for some societal groups than others while the feature selection part of the model development process (i.e., the selection of algorithm inputs) may be subject to the biases of the development team [4].
Beyond feature selection, AI algorithms are capable of demonstrating bias simply because of the way they are implemented. Predictive policing algorithms, for example, which may be used to allocate police resources across a city, have been shown to develop ‘feedback loops’ whereby police are repeatedly (and wrongly) sent back to the same neighbourhoods. It is believed that this behaviour occurs when such algorithms augment the results of its own police-deployment strategies to its existing set of training data; in other words, when decisions made by the system influence the data that the system will receive in the future
[12].
With many AI approaches, the learning process essentially involves tuning some parameters of the algorithm until an objective function is maximised. If the landscape associated with the objective function is complicated (i.e. possesses many local minima) then multiple training runs of the same approach can lead to very different results. This issue has been highlighted for a certain class of Deep RL algorithms [13], where sets of experimental trials were conducted using different random seeds (the resulting differences in algorithm performance arise because the initial parameter estimates of these algorithms are usually generated randomly). This lack of repeatability in the training procedure can obviously decrease the level of trust placed in the approach.
“Reproducing existing work and accurately judging the improvements offered by novel methods is vital to sustaining this progress. Unfortunately, reproducing results for state-of-the-art deep RL methods is seldom straightforward”
[Henderson, P., Islam, R., Bachman, P., Pineau, J., Precup, D., & Meger, D. (2018, April). Deep reinforcement learning that matters. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 32, No. 1).]
There is now a large range of open-source AI modules (see Python’s scikit-learn
[14], for example). These are often impressively large-scale projects and there is no doubt that they lower the cost-to-entry of many AI approaches, however, they also allow AI approaches to be applied with little understanding of their limitations. It is also worth noting that most of these AI modules are maintained entirely by volunteers; essentially, despite the use of detailed code testing procedures, there is no guarantee that this code is bug-free. Today (20th March 2021), for example, scikit-learn’s Github page features around 1,600 issues and 752 pull-requests (code contributions that are still being reviewed or developed). It is also worth noting that different implementations of what is nominally the same approach can also have a serious impact on algorithm performance
[13].
Clearly there many factors that can greatly influence the trust that is placed in AI algorithms. These same factors, as a result, significantly influence the level to which AI will be adopted within organisations that are interested in the technology.
In my experience, it is wrong to assume that addressing the main barriers to AI adoption requires the development of technical solutions (new algorithms, greater computer power etc.) While this, of course, important, the adoption process also needs to involve a great deal of education. AI solutions that appear to be opaque and are developed with limited input from the practitioners will often fail to be adopted, even if they are suitable for the task at hand. The key is to utilise an iterative development process, involving a close partnership between the AI team and practitioners, allowing the practitioners to learn about the reality of AI while also allowing the AI team to learn about the context where the algorithms will be applied.
Among other things, this type of collaborative approach is very rewarding! It’s always fascinating to learn about the interesting ways in which people want to use AI.
[1] Hall, W., & Pesenti, J. (2017). Growing the artificial intelligence industry in the UK. Department for Digital, Culture, Media & Sport and Department for Business, Energy & Industrial Strategy. Part of the Industrial Strategy UK and the Commonwealth.
[2]Ha, D., & Schmidhuber, J. (2018). World models. arXiv preprint arXiv:1803.10122.
[3] Alcorn, M. A., Li, Q., Gong, Z., Wang, C., Mai, L., Ku, W. S., & Nguyen, A. (2019). Strike (with) a pose: Neural networks are easily fooled by strange poses of familiar objects. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 4845-4854).
[4] https://www.gchq.gov.uk/pdfs/news/artificial-intelligence.pdf
[5] https://www.gov.uk/government/publications/guidelines-for-ai-procurement
[6] Marcus, G. (2020). The next decade in AI: four steps towards robust artificial intelligence. arXiv preprint arXiv:2002.06177.
[7] Koza, J. R., & Koza, J. R. (1992). Genetic programming: on the programming of computers by means of natural selection (Vol. 1). MIT press.
[8] Silver, D., Schrittwieser, J., Simonyan, K., Antonoglou, I., Huang, A., Guez, A., ... & Hassabis, D. (2017). Mastering the game of go without human knowledge. nature, 550(7676), 354-359.
[9] Vinyals, O., Babuschkin, I., Czarnecki, W. M., Mathieu, M., Dudzik, A., Chung, J., ... & Silver, D. (2019). Grandmaster level in StarCraft II using multi-agent reinforcement learning. Nature, 575(7782), 350-354.
[10] Ibarz, J., Tan, J., Finn, C., Kalakrishnan, M., Pastor, P., & Levine, S. (2021). How to train your robot with deep reinforcement learning: lessons we have learned. The International Journal of Robotics Research, 0278364920987859.
[11] https://www.gov.uk/government/publications/ai-roadmap
[12] Ensign, D., Friedler, S. A., Neville, S., Scheidegger, C., & Venkatasubramanian, S. (2018, January). Runaway feedback loops in predictive policing. In Conference on Fairness, Accountability and Transparency (pp. 160-171). PMLR.
[13] Henderson, P., Islam, R., Bachman, P., Pineau, J., Precup, D., & Meger, D. (2018, April). Deep reinforcement learning that matters. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 32, No. 1).