Published online by Cambridge University Press: 14 June 2025
Introduction
For a class of artificial intelligence (AI) systems, answers to the question of why – for some senses of ‘why’ – the system made a particular ‘decision’ are, as a matter of constitutive fact, unavailable. This discussion raises some issues that face Wittgenstein-inflected philosophers when thinking about this so-called Black Box problem. It offers reasons for thinking that Wittgensteinians should be intensely relaxed about such AI systems.
Black boxes
The black box problem arises because systems of the relevant kind are supposed to be ‘opaque’ to certain questions. I might ask, for example, why a system refused to give me a credit card (Bary n.d.), recommended a longer prison sentence, changed lanes on the motorway, offered good news about a cancer tumour (Gregory 2022), or determined a protein's shape (Trager 2022). Of course, if I were to query the system owners, I might, perhaps after some hesitancy, be answered: ‘Well, because our system determined that the target, you or your circumstances met (or failed to meet) criteria A, B and C’. The black box problem arises because, for a class of AI systems, this answer is false.
For the systems which do not interest us here – symbolic, procedural or GOFAI – this answer will be true. The input – the data about me and my circumstances – is an argument for the function here – which is calculated in this object here and so on. There is some value of a variable which, at some level of abstraction, is understood as the representation of some fact about you or your circumstances. Call these domain homogeneous systems. For such systems, someone can, in principle, run through the same inferential steps as the machine, with or without two kinds of pebble and a long enough toilet roll (Weizenbaum 1976, Ch 2). The result is rationally recoverable.
The AI systems which interest us here are convolutional neural networks with a primary discriminative (rather than generative) function. They are focused on finding a particular solution to a problem rather than modelling the data. Convolutional neural networks – ConvNets – are, we can say, domain heterogeneous. If someone asks why a certain result was forthcoming, we can describe how the system works, values of hyper-parameters, weights and activation biases, gradient descents, back-propagation algorithms and so.
To save this book to your Kindle, first ensure no-reply@cambridge.org is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Find out more about the Kindle Personal Document Service.
To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Dropbox.
To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Google Drive.