Safe and Trusted Artificial Intelligence 2021

post by Oliver Miles (2018 cohort)

Over three days from 12th-14th July 2021, I attended and participated in the Safe and Trusted Artificial Intelligence (STAI) summer school, hosted by Imperial College and Kings College London. Tutorials were given by leading academics, experts from British Telecom (BT) presented a session on industry applications, and I along with several other PhD students took part in a workshop speculating on AI interventions within the healthcare setting, presenting our work back to the wider group. In the following, I’ll summarise key contributors’ thoughts on what is meant by ‘safe and trusted’ in the context of AI and I’ll outline the themes and applications covered during the school I found to be most relevant to my own work. Two salient lessons for me expanded on contemporary efforts to reconcile accuracy with interpretability in models driving AI systems, and on efforts to systematically gauge human-human/human-machine alignment of values and norms, increasingly seen as critical to societal acceptance or rejection of autonomous systems.

When I read or hear the term ‘Artificial Intelligence’, even in the context of my peers’ research into present-day and familiar technologies such as collaborative robots or conversational agents, despite tangible examples in front of me I still seem to envision a future that leans toward science fiction. AI has always seemed to me to be intrinsically connected to simplistic, polarised visions of utopia or dystopia in which unity with some omnipotent, omniscient technology ultimately liberates or enslaves us. So, when it comes to considering STAI, I perhaps unsurprisingly default to ethical, moral, and philosophical standpoints of what a desirable future might look like. I obsess over a speculative AI’s apparent virtues and vices rather than considering the practical realities of how such futures are currently being realised and what my involvement in the process might mean for both me and the developing AI in question.

STAI began by addressing these big picture speculations as we considered the first theme – ethics of AI. According to AI professor Michael Rovatsos, ethical AI addresses the ‘public debate, impact, and human and social factors’ of technological developments, and the underlying values driving or maintaining interaction’ (2021). In a broad sense there was certainly agreement that ethical AI can and should be thought of as the management of a technology’s impact on contentious issues such as ‘…unemployment, inequality, (a sense of) humanity, racism, security, ‘evil genies’ (unintended consequences), ‘singularity’, ‘robot rights’ and so on (Rovatos, 2021).  An early challenge however was to consider ethics as itself an issue to be solved; a matter of finding agreement on processes and definitions as much as specific outcomes and grand narrative. In short, it felt like we were being challenged to consider ethical AI as simply…doing AI ethically! Think ‘ethics by design’, or perhaps in lay terms, pursuing a ‘means justified end’.

To illustrate this, if my guiding principles when creating an AI technology are present in the process as much as the end product, when I think of ‘safe’ AI; I might consider the extent to which my system gives ‘…assurance about its behavioural correctness’; and when I think of ‘trusted’ AI; I might consider the extent of human confidence in my system and its decision making’ (Luck, M. 2021). A distinction between means and end – or between process and goal – appeared subtle but important in these definitions: While ‘assurance’ or ‘confidence’ appear as end goals synonymous with safety and trustworthiness, they are intrinsically linked to processes of accuracy (behavioural correctness) and explicability (of its system and decision-making rationale).

In her tutorial linking explainability to trustworthiness, Dr Oana Cocarascu, lecturer in AI at King’s College London, gives an example of the inclination to exaggerate the trustworthiness in some types of data-driven modelling that ‘…while mathematically correct, are not human readable’ (Cocarascu, O). Morocho-Cayamcela et al. (2019) demonstrate this difficulty in reconciling accuracy with interpretability within the very processes critical to AI, creating a trade-off between fully attaining the two end goals in practice (Figure 1).

My first lesson for ‘doing AI ethically’ is therefore the imperative to demonstrate accuracy and explainability in tandem and without compromise to either. However, it doesn’t follow that this alone will ensure safe and trusted outcomes. A perfectly accurate and interpretable system may lead to confidence in mechanism, but what about confidence in an AI’s apparent agency?

In her tutorial ‘AI, norms and institutions’, Dr Nardine Osman talked about the ‘how’ of achieving STAI by means of harnessing values themselves. She convincingly demonstrated several approaches employing computational logic (e.g. ‘if-then’ rules) in decision making algorithms deployed to complex social systems. The following example shows values of freedom vs safety as contingent on behavioural norms in routine airport interactions expressed as a ‘norm net’ (Fig.2).

Serramia et al. visualise their linear approach to ethical decision making in autonomous systems, positioning conventionally qualitative phenomena – human values (e.g. safety) – as contingent on and supported by societal norms, e.g. of obligation to provide passports/forms (2018). Efforts to break down and operationalize abstract norms and values quantitatively (e.g. weighting by hypothetical preference, observed occurrence) demonstrate how apparent features of human agency such as situational discernment might become more commonplace in negotiating safe and trusted outcomes.  My second lesson and main takeaway from STAI’21 was therefore the imperative of sensitising AI, and design of AI, to the nuances of social values – distinguishing between value preferences, end-goals, social norms and so forth.

Lastly and significantly, attending and participating in STAI’21 has given me invaluable exposure to the practicalities of achieving desirable AI outcomes. The focus on ‘doing AI ethically’ has challenged me to pursue safety, trustworthiness, and other desirable qualities in my own work – mechanistically in terms of ensuring explainability of my methods and frameworks; and substantively, in terms of novel approaches to conceptualising values and positioning against social norms.


References

Cocarascu, O (2021) XAI/Explainable AI, Safe and Trusted AI Summer School, 2021 https://safeandtrustedai.org/events/xai-argument-mining/

Luck, M (2021), Introduction, Safe and Trusted AI Summer School, 2021 https://safeandtrustedai.org/event_category/summer-school-2021/

Morocho-Cayamcela, Manuel Eugenio & Lee, Haeyoung & Lim, Wansu. (2019). Machine Learning for 5G/B5G Mobile and Wireless Communications: Potential, Limitations, and Future Directions. IEEE Access. 7. 137184-137206. 10.1109/ACCESS.2019.2942390.

Osman, N (2021) AI, Norms and Institutions, Safe and Trusted AI Summer School, 2021 https://safeandtrustedai.org/events/norms-and-agent-institutions/

Rovatsos, M (2021) Ethics of AI, Safe and Trusted AI Summer School, 2021 https://safeandtrustedai.org/events/ethics-of-ai/

Serramia, M., Lopez-Sanchez, M., Rodriguez-Aguilar, J. A., Rodriguez, M., Wooldridge, M., Morales, J., & Ansotegui, C. (2018). Moral Values in Norm Decision Making. IFAAMAS, 9. www.ifaamas.org

Reflection of a Colloquium

post by Peter Boyes (2018 cohort)

As part of the programme with my industry partner Ordnance Survey (OS), each year I attend what they call a Research Workshop. It’s a multi-day trip down to their headquarters in Southampton, where they host all their sponsored PhD and Post-Doc students for a colloquium from their partner universities and programmes, both in the UK and a couple from abroad. The days consist of presentation sessions broken into themes of research, these presentations are given by each of the sponsored researchers to an audience of the other colloquium attendees and OS staff who drop in to relevant and interesting themes or talks over the days. In the breaks between presentation sessions there are poster sessions, each student presenting a poster of their work and able to talk with staff or other attendees there. These posters are also displayed over the course of the event to enable staff to drop by and take a look while they may be unable to attend a full presentation session, note questions and get in touch by email or later on in a break when the researcher is free. In addition there’s often a keynote speaker that kicks off the morning session talking around the general theme for each day.

As an annual event I have been able to attend at different stages of my PhD, and see progression across the visits. My view of the purpose of the event changed over appearances, and so did my confidence in my topic and myself. The conference-style event, presenting a poster, giving a talk, handling a Q&A with OS staff and fellow postgraduate researchers gave me a chance to learn from people going through the same process and some advice from them at their different stages of the postgraduate timeline. Over multiple poster sessions I honed the elevator pitch of my research for that year, and developed an understanding of my blind spots, the recurring questions that obviously I hadn’t anticipated or covered well enough in the poster, while developing my communication skills to multidisciplinary audiences. This was an opportunity to see others’ work that was similar to my field in different ways, and to practice communicating the research I was hoping to do or had done at the time of the workshop.

There is something to be said for not having any supervisors there, a little bit of a shock for me in my first year still settling into the doctoral training program at Nottingham. The student-supervisor relationship is a valuable one when navigating a PhD, but at this event I felt truly independent. At similar style events such as our Horizon CDT retreat I feel like even if they don’t contribute in my presentation, my supervisors are there in the background in the room or on the Teams call and may step in with comments or questions to jolt me along or help, but this wasn’t like that. This was more akin to what I expect conferences to feel like as I prepare to attend one and present later this year. Their contribution is there in the work, but I must be able to present and discuss the research as an independent researcher.

The event and this write-up gave me an exercise in reflecting on what stage I am at in my research. My first time attending, I was in the first year of the course, 5 months or so into my PhD and hadn’t exactly done an explicit research activity or carried out a study to talk about, I was still finding my feet. In that year, I talked mostly about my higher education background, my interests in a wide scope, essentially proposing questions I could explore and using the session to gauge some feedback on areas others thought could be interesting. This included areas to explore or advice on going down those paths, suggested literature or studies. Helpfully at this OS workshop there was an industry perspective on the applications and not just the theory or literature side or presentations.

In the next year, I could see for myself when making my presentation that my scope was narrowing, I was settling into an academic area, research questions were emerging less fuzzy, more defined even if not settled on at that point still. With the audience I was more engaged in discussion of conducted or planned studies and details of these, and looking towards potential research output goals and again the applicability to other sectors and industry.

With one of these trips to Southampton left to attend in my final run to thesis submission I will hopefully be in early write-up stages, and will be able to demonstrate some really interesting findings from this last year and my final study, and engage with those in their first years attending the workshop about their experiences in the PhD journey to that point.

To bring this to a conclusion, I would encourage postgraduate research to look for these colloquiums/consortiums even if not offered by your industry partner as they can help you engage with your research in a different way. These are an opportunity to participate without the same pressure or work of preparing a paper and submitting to a journal or conference, those are different experiences, both highly beneficial. I would also recommend in the way writing this has been for me, to engage with reflective exercises for your journey to recognise, even if for just yourself, the work you have been doing, the changes and narrowing of scope, and your understanding of a field or concepts. I would also encourage industry partners with multiple postgraduates across the country to try and organise events like these to support their development, and help to establish academic and industry networks they may be struggling with confidence or opportunities to build beyond their own centre or institution.