Director Essentials
Implementing AI Governance
Learn how boards can oversee AI strategy, risk, and technology competency.
Governance Surveys
Directorship Magazine
Director Essentials
About this Report
Artificial intelligence (AI) applications are already impacting business and society. It’s time for boards to oversee AI’s unique risks and help their firms to seize the opportunities that AI presents. This edition of Director Essentials outlines the boardroom’s role in AI governance and helps directors become conversant in the basics of the underlying technologies. It then introduces the operational and regulatory matters that can shape a company’s AI journey.
How Boards Can Use This Resource
Artificial intelligence (AI) is a transformational technology with the potential to produce both significant value and harm for businesses and their stakeholders. With advances occurring rapidly amid widespread attention, it can be difficult for a company to cut through the hype to understand whether and how this technology can positively impact its business. However, most companies are likely already using some forms of AI in their operations through tools the organization has acquired or developed, or from services provided by suppliers.
AI poses unique opportunities and risks, both of which implicate boards’ fiduciary responsibilities. Boards can continue to draw on the suite of governance principles, IT governance frameworks, practices, and experiences honed through their oversight of cybersecurity, privacy, ethics and compliance, and emerging technology to meet the challenges of AI governance.
EMERGING TECHNOLOGY GOVERNANCE PRINCIPLES
On the other hand, artificial intelligence does require the board to shift its perspective, its engagement with management, and some of its practices. This publication outlines foundational AI knowledge, presents key board focus areas, and describes how the board can adapt to provide effective oversight. Boards will likely need to update committee responsibilities, increase AI-specific education, and engage more frequently and in greater depth with management in its evaluation of AI.
AI presents a challenge to traditional forms of governance due to its unique capabilities and the environment in which it is being developed. Once trained, AI systems can develop rules and propose choices without human input. As a result, traditional governance practices that are designed for oversight of technologies and decision-making processes that center on human judgment and use are not always fit for purpose.
Further complicating the challenge for boards as they oversee and govern AI use in their companies are three aspects of today’s AI systems:
PREPARING FOR AI GOVERNANCE
NACD 2023 survey data reveals that 95 percent of directors believe AI will impact their business in some way in the next year. However, only 28 percent of respondents indicate that AI is a regular feature in board conversations. With boards anticipating a significant impact from AI on their businesses, boards must evaluate their own readiness and prepare themselves for this new oversight area in the five areas indicated below. After that, deeper learning is required, as described in this guide, concerning the technical concepts involved in sound management of AI.
| Impacts on Strategy and Risk | Understand the company’s current engagement with and use of AI and how to integrate AI into discussions of strategy and risk. |
| Committee Responsibility | Leverage the strengths of existing committees to oversee the various components of AI where they would be most appropriately governed. |
| Board Composition | Evaluate board composition in context of changes stemming from AI. The technology may warrant new skills, experiences, and backgrounds on the board to provide effective oversight. |
| Director Education | Invest in onging director-specific education, including management presentations, outside experts, personal engagement with AI research, or attendance at AI events. |
| Management Reporting | Work with management to receive much more frequent, forward-looking updates about the company’s AI initiatives. |
Stanford professor John McCarthy coined the term artificial intelligence in 1955. Since then, the field has evolved in multiple ways, propelled by advances in data collection and storage, chip design, and data analytics. The technology’s most significant advances have been driven by “machine learning” systems, which are not programmed but are trained on data sets and then continuously adapt. They “learn” by analyzing more data. This architecture, based on arrays of “neural networks” modeled on the neurons and synapses of the brain, is called “deep learning” (“deep” because it contains multiple layers of such neural networks).
WHAT ARE AI MODELS?
An AI model is “a learned representation of underlying patterns and relationships in data, created by applying an AI algorithm to a training data set. The model can then be used to make predictions or perform tasks on new, unseen data.” Advances in machine learning systems have created models able to independently parse data, learn, and then make informed predictions based on what the model learned from the analyzed data. These models include the large language models (LLMs), which are also called “foundational” or “generative” models. Those models emerged at companies like Open AI, Deep Mind, Anthropic, Meta, IBM, and others.
HOW DOES AI WORK?
The foundation of AI performance is data, often from multimodal sources like text, speech, and images. The more and higher-quality data a system is fed, the better the AI. AI systems can only create as much value or be as trustworthy as the underlying data they are trained on and interact with while in operation.
The use of deep learning with artificial neural networks is what currently gives foundational and generative AI models like ChatGPT their ability to produce content comparable to that created by humans, across a broad range of tasks. This technique also allows models to improve and scale at much greater speeds. The rapid improvement in AI capabilities and the continued accumulation of large amounts of data mean that performance can improve at a near-exponential scale. AI learns at the speed of data. This also means that models can quickly become obsolete that just months ago were best in class. This speed and scale of AI systems' evolution both improves model performance across a range of applications and increases exposure to model risks and harms.
Many companies have already begun to adopt artificial intelligence and integrate it into their workflows and strategies. Whether your company is at the beginning of its journey or further down the path, board members should understand the key areas of AI opportunity and risk, along with the fundamentals for overseeing the technology.
STRATEGIC BENEFITS AND OPPORTUNITIES
As with most new and evolving technologies, AI presents significant opportunities to explore. It is imperative that directors understand these benefits, begin a dialogue about them with their management teams, and specifically ask questions about how, or if, AI can be a strategic asset to the company.
Directors who were in management during the rise of the Internet are likely to draw parallels between today’s AI and Web 1.0. Boards must approach AI oversight as an organization-wide strategic imperative and not simply as an IT operational issue or focus exclusively on risks. It is a common pitfall in technology governance to lose sight of potential value to the customer or other stakeholders. This means that while directors do not need a deep understanding of AI operations, boards should always seek to understand how AI can generate value for the business and the customer and orient discussions on the topic to that end. This will spark a much more productive conversation about AI between management and the full board and can help the company proactively adopt beneficial AI tools as strategic enablers before they become strategic threats.
COMMON APPLICATIONS FOR AI
RISK & LIABILITY MITIGATION
It is the board’s role to ensure that management is appropriately monitoring and mitigating the risks and harms in AI models. Algorithmic harm can be introduced via poor training-data quality, technical flaws in the algorithm, human error in development, or interacting with the algorithm outside intended use cases. Failure to address AI risks can result in further regulatory, reputational, or legal risks down the road, as well as failing to produce the expected value for the business.
It is important for boards to understand that managing risk in AI tools entails difficult decisions that will require tradeoffs. For example, optimizing a tool with privacy-enhancing features may sacrifice transparency about how the algorithm’s explainability and transparency work, which could impact compliance with certain trustworthy or responsible AI standards. These are important decisions because the AI model does not have any context or prior understanding of the world in which it exists. Individuals who are impacted by the decisions generated by AI may then be vulnerable to outputs generated in unjust, unexplainable, and harmful ways. For example, a model may not have been trained to understand the concept of human or civil rights, and thus, the model may inadvertently discriminate against certain classes of people.
Boards as AI stakeholders must be prepared to ensure that the development and use of AI remains safe and aligned to the company’s core values, ethics, purpose, and mission, all while taking the appropriate measures to mitigate harm. The organization’s values and ethics should be documented and regularly referenced by the board when engaging with management in oversight of AI.
COMMON AI DEVELOPMENT AND USE RISKS
i. Bias: During the process of developing and testing an AI model, the assumptions, preferences, and historical biases that exist in the training data or among developers can become embedded within the algorithm. This could lead to the production of discriminatory outputs by the AI.
ii. Explainability & Interpretability: Because machine learning systems learn from their own experience, rather than being preprogrammed, models can be “black boxes,” meaning that it is not always possible to understand and explain how the inputs and operations within a model produced an output. Lack of explainability prevents users from challenging AI-influenced outcomes and could expose the company to litigation and reputational risk.
iii. Privacy: AI models can jeopardize individuals’ privacy and introduce legal risk if insufficient care is taken to implement legally required privacy protections for the data used to train or operate the model.
iv. Security and Resilience: AI that is entrusted to support or control critical systems—transaction processing, supply chain, financial operations, etc.—introduce new vulnerabilities to those systems if the AI fails, is hacked, or has insufficient human oversight. Harm can also be caused to individuals via exposure of their data from a vulnerability or breach in the AI system. AI models also suffer from their own vulnerabilities, such as prompt injection attacks, and may require new controls to protect the model's integrity and performance.
v. Misuse: Once AI models have been released for public use there is the potential for misuse of the tool for malicious activities. Models have been used for activities like malware development, phishing campaigns, and the creation of disinformation and fake news. Models can also be used to produce untrue, synthetic data that can be fed back into the algorithm, which can then poison the model and render it ineffective and inoperable.
vi. Model Drift: AI models continue to evolve and change as they are used. If not monitored effectively, models can begin performing outside their intended use which heightens the risk for stakeholder, reputational, and legal harm.
Like digital transformation, AI impacts the entire organization and therefore requires a holistic approach to governance and oversight. Board governance of AI operations involves directors asking the right questions to ensure that the use and development of AI tools and systems remains aligned with the company’s overall mission, purpose, values, and ethics. Effective AI governance within the organization will blend strategy, corporate ethics, and values together with policies, procedures, and controls that guide responsible model development and use in the organization.
DATA GOVERNANCE & QUALITY
Data governance and quality assurance have long been established practices in many companies. However, their foundational importance takes on new significance with AI because models can operate and create results without human intervention. Thus, the performance, trustworthiness, and safety of the AI model is directly correlated with the quality of data on which it is trained. An organization is only able to develop trustworthy AI if the data used in training and operation of the model is of a sufficiently high quality. In fact, tasks related to data preparation and selection such as cleaning, transforming, and splitting will likely require the majority of time in model development and training. Data quality is a complex topic. Boards should focus their questions to management on accuracy, completeness, consistency, uniqueness, provenance, and fitness for purpose.
The board should request that management define data quality standards that meet the organization’s needs, along with metrics and information to validate alignment of the data used in AI systems with these standards. The board should also ensure that structures and processes are in place to prepare and validate the quality of both internal and external data sources before they are used in AI models.
CAPITAL ALLOCATION
AI model development or tool procurement can represent a significant capital expenditure. Licensing or acquiring finalized, purpose-specific AI tools can be economical but may be less purpose specific, while developing, training, and deploying original AI models is expensive but could lead to tailored outcomes.
Boards should also recognize that capital allocation with AI is not limited to technology investments but has enterprise-wide scope. For instance, boards must ask management how much the organization may need to spend on workforce retraining or increased compliance costs—two nontechnical, significant investments required to make the application of AI successful. One area where the board can update its processes is M&A transactions. Where AI represents a significant component of a target company’s business model, boards should insist that the due diligence process incorporate evaluations of AI value drivers and risk factors—from algorithmic diligence and data diligence to cultural diligence.
Scrutinizing capital allocations is a core governance function, and boards should subject the expenses tied to AI to the same rigor as other capital investments. And, given the high cost of model training, scrutiny of AI expenses can be an opportunity to ask questions about the broader governance of AI development and deployment within the company. A lack of good governance and development practices will lead to wasted capital on an unsafe AI model that must be retrained or abandoned, leading to loss on investment or a prolonged timeline to return on investment.
AI GOVERNANCE AND MANAGEMENT FRAMEWORK
AI oversight can be adapted from existing frameworks that have worked across other functions. Common models that can be adapted for AI governance include the Three Lines Model, which is frequently used in audit and cybersecurity, or a Hub and Spoke model. Whatever framework is chosen, it should be coupled with new organizational structures, technical processes, and controls that direct the use and development of AI within the organization. Typical practices include the use of an AI impact assessment prior to and during model development or AI tool implementation, as well as periodic algorithmic audits to certify that the model remains safe.
The board of directors should validate whether management has a governance framework with accountability and controls implemented within the organization. By asking questions and validating the framework, the board will be directing the organization toward responsible engagement with AI with appropriate management and accountability.
ETHICAL AND TRUSTWORTHY AI USE AND DEVELOPMENT
Oversight of corporate ethics and compliance is a core fiduciary duty of the board. Thus, the board has an obligation to oversee and validate whether AI is used and developed in a way that aligns with the characteristics of trustworthy AI. Trustworthy AI systems include those that are “valid and reliable, safe, secure and resilient, accountable and transparent, explainable and interpretable, privacy-enhanced, and fair with harmful bias managed.” Several responsible and trustworthy AI frameworks have been developed, with many consistent principles across frameworks.
The board should direct management to develop, acquire, or implement only those AI systems that include the principles and characteristics of trustworthy AI and that also align with the company’s ethics and values. The goal of governance in the responsible development of safe systems is to instill trust, deliver benefits to the business, reduce harm to the greatest extent possible to all stakeholders, and align to the company’s core values, strategy, mission, and purpose.
IMPACTS ON WORKFORCE
AI has the potential to change business models, automate entire functions within the organization, and augment the tasks performed by a majority of a company’s employees. Research in this area has revealed that AI can significantly impact work tasks of up to 80 percent of US employees. Likewise, IBM research estimates that “more than 120 million workers in the world’s 12 largest economies may need to be retrained or reskilled.” Boards should request information from management on AI’s impact on the company’s employees while also asking questions about how AI could affect the company’s workforce size; recruiting; efforts around diversity, equity, and inclusion; and costs for retraining and skill development.
Research has shown that many HR functions are already using AI for screening and recommending job candidates, recommending compensation and employee learning plans, and predicting attrition rates. Consequently, boards should inquire about the degree to which the company’s HR function uses these systems and then update their human capital oversight practices to account for how AI is impacting the workforce. The board should also evaluate the organization’s current workforce and leadership to assess their readiness for AI-driven organizational change and consider how AI may impact the company’s future workforce needs for long-term success.
AI REGULATORY COMPLIANCE
AI regulation at present is nascent and highly fragmented. Significant regulations include New York City’s recently passed law that will require companies to independently audit any AI tools that are used in hiring decisions. Similar laws are proposed in New Jersey, California, Vermont, and Washington, DC.
The European Union’s proposed AI Act, a law similar in scope and potential impact to the EU General Data Protection Regulation, proposes a classification system for AI tools with a variety of requirements and obligations tailored to a risk-based approach. Tools would be authorized in the EU market only if they meet the requirements outlined in the proposed law. Parallel to these regulations, some companies are adopting voluntary commitments for the safe and responsible development and use of AI where no clear, national regulatory framework exists, such as the United States.
Even though the regulatory environment for AI remains complex and emergent, boards should take a proactive approach and strive to remain ahead of AI regulatory requirements. The board should work with management to anticipate the impact of upcoming AI regulations on the business, with the understanding that this space is constantly evolving and tends to eventually align with what emerges from the European Union. Compliance failures represent a significant risk, and boards have a responsibility to oversee the company’s compliance function. Fulfilling this responsibility for AI regulation will be challenging, but regulations may become a lever to ensure that companies are engaging with AI systems safely and responsibly.
AI REGULATORY APPROACHES TO MONITOR
As regulations develop, it can be helpful for the board to assign responsibility for tracking AI regulatory matters to a chief legal officer or general counsel.
Which business units and functions use AI tools?
In what way are the tools used?
What type of tools are being used? Are we building them, or were they procured?
Is our management team aware of how and when employees are using publicly available AI tools and models to complete tasks?
What use cases are we exploring?
How does AI impact the company’s strategy and approach to risk?
What organizational gaps hinder the development of AI that may give us a strategic advantage?
What are AI’s potential impacts on company strategy, business model, and industry?
Who within the organization should report to the board on AI matters? Do we have a consensus on metrics and information the board would like to see regarding AI initiatives?
How often does the full board need to be reported to on these matters?
Which committee will perform primary oversight of AI, if any?
Data quality
Capital allocation
AI governance and management framework
Ethics oversight and responsible development
Workforce impacts
Regulatory compliance
Are investments in AI-specific education for the board are necessary?
Are board structures and processes appropriate to ensure responsibility for AI?
Have we delegated responsibility to a committee to oversee this issue?
Which components of AI oversight are best handled by different committees? (e.g., risk, algorithmic auditing, workforce impacts, regulatory and compliance oversight, etc.).
Do committee charters require updates to reflect these new responsibilities?
Registration Now Open for
NACD Directors Summit™ 2026
Register Early and Save $2,000
October 11-14, 2026
The Gaylord National Harbor | Washington, DC Area