Article

Insights from Practice: Telefónica’s AI governance journey

Artificial intelligence is everywhere. It improves our business (semi-automated decisions, chatbots, fraud detection, transport and logistics, autonomous vehicles), improves our daily lives (spam filtering, search engines, machine translation, recommendations, optimized routes), and can help run better societies (breast cancer diagnosis, diabetic eye disease, education, school dropout detection, increasing responsiveness to calamities, understanding effects of climate change), to name just a few examples.
Telefónica Logo

Authors:  Richard Benjamins, Chief Responsible AI Officer, Joaquina Salado, Head of AI Ethics, Violeta Albuquerque, AI Ethics and User Experience Designer 

However, the massive adoption of AI also involves ethical, social, and societal risks and challenges. Who hasn’t heard about AI systems that discriminate, are black boxes, have too much autonomy, take our jobs, increase fake news, invade privacy, ignore intellectual property rights, and interfere with democratic processes?  

In general, there are three approaches that we today use to deal with those risks: 1) regulation (e.g., the European Union’s AI Act, the US Executive Order on Safe, Secure & Trustworthy AI, the draft Brazilian AI Act, Chinese regulations on generative AI), 2) international recommendations (UNESCO, OECD), and 3) self-regulation.  

In this brief article, we discuss the journey of a large telecommunications company, Telefónica, going from the publication of high-level ethical AI principles (in 2018) to an approved internal AI governance model (in 2023). The figure below illustrates this journey. Over the past 10 years, Telefónica has implemented more than 500 AI applications across many different business areas. For instance, we use AI to improve the quality of end-customer services by optimizing our network and its security; as well as to personalize customer-oriented offers in order to generate a better experience. 

Adopting ethical AI Principles (2018) 

The first step of our journey was to come up with a set of ethical AI principles. In 2018, there were not many ethical AI initiatives that we could choose from, so we investigated with our experts and came up with a company-wide agreed set of principles in line with our Responsible Business Principles. Today, it is much easier since company can decide to adhere to existing initiatives such as UNESCO’s Recommendation on the Ethics of AI or the OECD’s AI Principles. The specific principles are shown in the figure below. We are currently in the process of adding a new principle related to “Green AI” which relates to minimizing the emission footprint of AI systems.  

Designing a methodology: Responsible AI by Design (2019) 

Implementing AI principles in an organization requires a methodology. We have designed the “Responsible use of AI by Design” methodology to facilitate this process. The main ingredients of this methodology are: 

  • The AI principles 
  • An awareness and training program 
  • A questionnaire to assess the different AI applications 
  • Technical tools such as for explainability, fairness and environmental impact. 
  • A governance model  

The first step to bring the methodology to life was to create an online course on AI Ethics, explaining the basics of AI along with basic ethical questions that may arise when using the technology. So far, around 4000 employees have finished the course.  

Experimenting with how to implement the methodology (2020-2021) 

The second step was creating a questionnaire to facilitate performing ethical assessments of the products and services of the company. The first version consisted of a small set of question for each ethical principle. In a second version, we added a risk assessment phase where we asked product owners to assess the impact in case the AI system made mistakes in terms of severity, scale, and probability of occurrence. In collaboration with the Environment, Society, and Governance (ESG) department, this version was trialed in one of our main business operations in Spain by including it in the formal product development process. In a third version, we also included the future requirements of the European Union’s AI regulation, the AI Act.  

Running pilots to confirm and adjust the methodology (2022) 

Once we confirmed the questionnaire from a content and usability perspective, we designed an AI governance model that we wanted to test in several business units. Our governance model introduced three new organizational roles: 

  • Responsible AI (RAI) Champions: a diverse and multidisciplinary group in charge of fostering the responsible use of AI in the business areas they represent, providing ethical guidance to the product owners, proposing measures to mitigate possible risks, and ensuring training and awareness raising on the topic.  
  • An AI Ethics “Committee”: a multidisciplinary group of experts in AI where ethical risks can be escalated for guidance and/or decision-making.  
  • An AI “Office”: a small team responsible for leading the change management process needed for the responsible use of AI in the business units where AI systems are developed, used, acquired, or marketed. The Office coordinates and/or executes the necessary risk assessments, and monitors the implementation of the requirements. The team also coordinates the activity of the RAI Champions and the AI Ethics Committee.  

The governance model also introduced a workflow, and assigned responsibilities in the workflow to different roles and/or departments, such as the product owner, the RAI champion, and the ethics committee. Moreover, we designed a problem-based escalation process. AI use cases that were assessed as high-risk, or had disagreements in the assessments, were escalated to the ethics committee for debate and recommendations. If the debate was not conclusive, final decisions were supposed to be taken by the Office of Responsible Business. From this point, we strengthened our collaboration with the Compliance department via weekly meetings.  

After running pilots in four different global business units (Innovation, Human Resources, Enterprises, CDO), we extracted many learnings, some of which include: 

  • The importance of having a clear and well-known definition of AI. 
  • The need to identify new essential roles to drive the cultural change implied in the process, such as the RAI Champions, the Ethical AI Committee, and an “AI Office” that drives the entire program. 
  • The AI value chain is more complex than simply providers and users, and a thorough review of the different types of roles a company can have and the responsibility it entails is necessary. The roles include AI providers, deployers, partners, buyers, etc.  
  • Foster a culture of responsible design for instance by creating guides and training programs for the different profiles in the AI value chain. This culture has motivated employees working with AI to develop a sensibility for ethical thinking; to develop technical tools to mitigate specific risks (bias, black box algorithms); and bring the consumer voice into our AI product development processes.  

Approving an internal AI governance regulation (2023) 

Equipped with the practical experience of the governance model obtained in the pilots, and in close collaboration with the Compliance area, we created the official AI governance model of Telefónica which was approved in December 2023. This final governance model also includes the requirements stemming from the European Union’s AI Act, UNESCO’s Recommendation on the Ethics of AI, and also from some other relevant initiatives such as the US National Institute of Standards and Technology (NIST) and the OECD. 

Rolling out the governance model across the Group (2024) 

2024 will be the year of rolling out the AI governance model across the Group in a collaboration between the Compliance area and the AI Office, and supported by an application for 1) registering AI systems (performed by the business units in collaboration with the RAI Champions), 2) assessing the risks (performed by the AI Office), 3) identifying corresponding risk-mitigating requirements (legal, security, ESG, privacy, etc.), and 4) implementing the requirements (business units). The application will provide clear and robust accountability. In case ethical doubts appear in the process, the ethical AI committee will provide recommendations which - when needed - can be escalated to the maximum level.  

Benefits of the responsible use of AI 

We believe that there are several benefits for a company of using AI in a responsible manner, including: 

  • Increasing the company value for investors 
  • Increasing customer trust and reputation 
  • Business benefits 
  • New B2B business opportunities 
  • Attracting and retaining talent  
  • Better positioning to governments 

The following figure provides more detail of each of the mentioned benefits. 


The ideas and opinions expressed in this article are those of the author and do not necessarily represent the views of UNESCO. The designations employed and the presentation of material throughout the publication do not imply the expression of any opinion whatsoever on the part of UNESCO concerning the legal status of any country, city or area or of its authorities, or concerning its frontiers or boundaries.