TLDR
Only 12% of C-suite executives can confidently evaluate an AI-generated output for accuracy and appropriateness without assistance — yet 81% say AI is a top-three strategic priority for their organisation in 2026, according to research from PwC's Global CEO Survey. This guide defines what AI literacy actually means at the executive level, introduces the AI Leadership Capability Model (ALCM) with its four-quadrant self-assessment, and provides a 90-day literacy sprint that has moved leadership teams from average ALCM scores of 38 to 71 out of 100.
Contents
- The Executive AI Knowledge Gap in 2026
- What AI Literacy Actually Means for Senior Leaders
- Strategic Evaluation: Knowing Where AI Creates Real Value
- Team Capability Judgment: Assessing What Your People Can Actually Do
- Investment and Vendor Literacy: How to Evaluate AI Without Being Misled
- Governance and Risk: What Every Executive Must Have in Place
- The AI Leadership Capability Model (ALCM): A Self-Assessment
- Case Study: Global Consulting Firm, 120 Partners
- Measuring Progress: The AI Literacy Index for Leadership Teams
- Getting Started: A 90-Day Literacy Sprint for Executives
The Executive AI Knowledge Gap in 2026
The gap between executive rhetoric on AI and executive capability in AI is now one of the most consequential mismatches in corporate governance. Boards and leadership teams are approving significant AI investments — in tools, platforms, training, and infrastructure — while lacking the foundational literacy to evaluate whether those investments are well-designed, whether vendors are credible, whether their teams have the skills to execute, or whether the risks have been adequately addressed.
The PwC Global CEO Survey 2025 found that 81% of CEOs describe AI as one of their organisation's top-three strategic priorities. The same survey found that fewer than one in eight C-suite executives could demonstrate confidence in evaluating the accuracy and appropriateness of AI-generated content without assistance. This is not a peripheral skill gap. It is the primary strategic risk in AI adoption: leaders who cannot evaluate AI outputs, AI vendors, or AI capability claims are making decisions in a fog.
The reasons for this gap are structural. Most senior leaders reached their current positions through mastery of domains — finance, law, operations, consulting, technology — in which AI, in its current form, simply did not exist. Their mental models of technology capability were formed in an era when meaningful software required engineers. Their instincts about what is feasible in a given timeframe, at a given cost, for a given level of quality, are calibrated to a pre-AI world. These instincts are now systematically wrong.
This is not a criticism of the leaders themselves. It is a description of a structural obsolescence problem that affects virtually every leadership team in every sector. The question is not whether the gap exists — it does, nearly universally — but how quickly and effectively it can be closed.
AI literacy for executives differs from AI training for their teams in two important respects. First, the goal is not operational proficiency. A CEO does not need to build AI tools; they need to be able to evaluate them. A CFO does not need to know how to prompt an LLM; they need to be able to assess whether their finance team's AI tools are producing reliable outputs and whether the investment in those tools is generating a return. Second, the format of development that works for executives is fundamentally different from the format that works for individual contributors. Lecture-based programmes, while often used for executives, are among the least effective approaches. Demonstration, structured scenario practice, and peer learning work significantly better.
What AI Literacy Actually Means for Senior Leaders
The most common misconception about AI literacy at the executive level is that it means having a working understanding of the technology. Some programmes respond to this misconception by teaching executives how transformer models work, how AI training data is constructed, or how to differentiate between various categories of machine learning. This is technical literacy. It is not executive AI literacy. The two are different things, and conflating them produces the wrong development outcomes.
Executive AI literacy is the set of capabilities that allow a senior leader to make better decisions in four domains: where AI creates genuine value in their organisation (strategic evaluation), what their teams can and cannot currently do with AI (team capability judgment), which AI investments and vendors are credible (investment and vendor literacy), and what risks require governance attention and what minimum safeguards look like (governance and risk). These are the decisions that matter at the executive level, and they require a specific form of literacy — not technical depth, but informed practical judgment.
A useful analogy is financial literacy. We do not expect senior leaders to be able to produce a set of management accounts or conduct a DCF analysis. We do expect them to be able to read a P&L, identify anomalies, ask the right questions of their finance team, and make sound decisions about financial resource allocation. The equivalent standard for AI literacy is precisely calibrated: senior leaders should be able to evaluate an AI proposal, assess whether a vendor claim is credible, understand what a capability assessment is telling them about their team, and identify the governance gaps that require attention before they materialise as risks.
The AI Leadership Capability Model (ALCM) defines this standard across four quadrants. Each quadrant corresponds to one of the four decision domains, and each carries a self-assessment score and an action guide. The model is designed to be diagnostic — it tells leaders not just where they are, but specifically what development would move them forward in each quadrant.
Strategic Evaluation: Knowing Where AI Creates Real Value
Strategic evaluation is the ability to assess, with reasonable accuracy, whether a proposed AI application will create genuine, measurable value for the organisation — as distinct from AI applications that are technically feasible but commercially irrelevant, or AI applications that create the appearance of progress without underlying value creation.
Most senior leaders, when presented with an AI proposal, lack a systematic framework for this evaluation. They respond to confidence, enthusiasm, and technical fluency in the presenter rather than to the quality of the underlying value case. The result is AI investment that tracks what is demonstrable in a pitch rather than what is valuable in production.
Effective strategic evaluation requires three capabilities. The first is value chain analysis: the ability to identify, in any given business function, which tasks are characterised by high volume, structured inputs, and rule-based processing — the categories of work where AI creates the most reliable value — and which tasks are characterised by novel inputs, complex judgment, and high-stakes client interaction, where AI's contribution is more limited and more closely supervised. Leaders who cannot perform this analysis at a basic level will consistently misjudge where AI investment should flow.
The second capability is feasibility calibration: an accurate mental model of what AI can and cannot do reliably in a production context, as distinct from a demo context. AI tools in demonstration produce more reliable outputs than in production, because demos are designed around the tool's strengths. A leader who calibrates AI's capabilities from demo performance will consistently overestimate what a production deployment will achieve. Feasibility calibration comes from exposure — specifically, from watching AI tools work on real tasks in real conditions, including watching them fail and understanding why.
The third capability is timeline realism: understanding how long it takes to move from an AI tool deployment to a measurable productivity outcome, accounting for the capability development, workflow redesign, calibration, and change management that responsible implementation requires. Organisations that expect AI tools to produce ROI within 30 days of deployment are setting themselves up for disappointment. The realistic timeline from first deployment to first measurable outcome, for a well-designed programme, is 90-180 days.
Team Capability Judgment: Assessing What Your People Can Actually Do
One of the most significant consequences of the executive AI knowledge gap is the inability of senior leaders to make accurate judgments about their teams' AI capability. Leaders who cannot assess AI capability themselves are entirely dependent on what their teams tell them — and in an environment where AI capability has become a professional currency, self-reports are systematically biased upward.
This creates a specific problem at the strategy level. Senior leaders making decisions about AI-dependent projects and programmes need an accurate picture of whether their teams can execute. A digital transformation programme that assumes AI-capable analysts, deployed into an organisation where most analysts are at Stage 1 or Stage 2 on the AI Capability Matrix (see the AI Skills Gap Analysis guide), will fail to deliver its projected value — not because the strategy is wrong, but because the capability assumption was incorrect.
Team capability judgment at the executive level does not require the leader to conduct skills assessments personally. It requires them to know what a credible capability assessment looks like, what questions to ask of an assessment result, and how to read the output in a way that informs strategic decisions.
The key questions a senior leader should be able to ask of any AI capability assessment presented to them are:
- What was the assessment methodology — self-report only, or observed performance?
- What is the distribution of scores across the three dimensions (Literacy, Integration, Construction), and what does this tell us about the nature of the gap?
- Which roles and functions are most materially below target, and what is the business consequence of that gap?
- What training investment would be required to close the priority gaps, and over what timeframe?
- How does our team's AI capability compare to industry benchmarks or to what we know of our direct competitors?
A senior leader who can ask and evaluate the answers to these questions is exercising genuine team capability judgment. One who simply accepts a summary score without interrogating its basis is not — and is likely to make resource allocation decisions that are not grounded in the actual capability of their people.
Investment and Vendor Literacy: How to Evaluate AI Without Being Misled
The AI vendor market in 2026 is characterised by a combination of genuine capability, inflated claims, and a significant gap between demo performance and production performance. Senior leaders who lack vendor literacy are systematically disadvantaged in procurement decisions. They cannot distinguish between vendors with substantive, tested solutions and vendors with compelling pitch decks and technically impressive demonstrations that do not hold up in production.
Investment literacy means being able to ask the questions that reveal whether a vendor's claims are grounded. The ten questions that distinguish credible AI vendors from those whose value is primarily in the sales process:
- Which specific tasks does your product perform reliably in a production environment, not a demo? What are the known failure modes?
- What is the average onboarding-to-first-measurable-outcome timeline for a client comparable to us in size and function?
- What does implementation require from our team — in time, in technical resource, and in change management?
- How do you handle data security for inputs that contain confidential client or employee information?
- What does your SLA look like for accuracy? And what happens when the output is wrong?
- Which of your existing clients would describe you as a production partner rather than a proof-of-concept vendor? Can we speak to them?
- What capability does our team need to have before deployment to realise the value you're projecting?
- How does your pricing structure change as usage scales? Are there volume thresholds that materially change the unit economics?
- What is your model for product development in the next 18 months, and how do your current clients influence it?
- If this deployment underperforms against the projected outcomes, what is the exit mechanism and what are the costs?
Vendors with genuine, production-tested products will answer these questions directly and specifically. Vendors whose value is primarily in the demo will deflect, generalise, or redirect to the pitch deck. The pattern of responses is itself informative.
Investment literacy also extends to evaluating build-versus-buy decisions: the judgment about when to purchase an AI solution versus when to develop a custom capability using general-purpose AI tools. This decision is often made on the basis of what is visually compelling in a vendor presentation rather than on the basis of total cost of ownership, strategic fit, and organisational capability to integrate and maintain a purchased solution. Senior leaders with strong investment literacy can evaluate this decision on its merits.
Governance and Risk: What Every Executive Must Have in Place
Governance literacy at the executive level is not about understanding every detail of AI regulation or data protection law. It is about knowing the five things that must be in place before AI is deployed at scale in the organisation, and being able to identify quickly when one of those five things is absent.
1. A data classification policy for AI inputs. The organisation has defined which categories of data can be used as inputs to AI tools, which cannot, and under what conditions sensitive data can be used with appropriate safeguards. Without this, individuals across the organisation are making their own decisions about what to feed AI tools — creating data exposure risks that the organisation cannot monitor or manage.
2. A quality control standard for AI-assisted outputs. The organisation has defined the minimum review requirements for AI-generated content before it is used, shared, or acted upon. The standard varies by output type and risk level: a marketing email has a different quality control requirement from a regulatory filing or a client investment recommendation. But the standard is documented and consistent — it is not left to individual judgment.
3. An AI use disclosure framework. For external-facing work, the organisation has a clear position on when and how AI use is disclosed to clients, customers, or counterparties. This framework addresses both regulatory requirements (which are evolving rapidly) and client expectations, and it is applied consistently rather than case-by-case.
4. A vendor due diligence protocol. Before any new AI tool is deployed, a defined due diligence process is conducted covering data security, model behaviour, vendor financial stability, and contractual protections. The protocol prevents the pattern — common in many organisations — where AI tools are adopted by individual teams or functions without IT, legal, or compliance review.
5. An accountability structure. It is clear who in the organisation is accountable for AI governance overall, and who is accountable for AI use within each function or business unit. The accountability structure does not require a Chief AI Officer if the organisation is not large enough to justify one — but it does require someone whose role explicitly includes AI governance oversight and who has the standing to escalate concerns to the executive team.
An executive who can confirm all five of these are in place has a functioning AI governance baseline. An executive who cannot confirm all five has governance gaps — and should treat those gaps as an urgent priority, not a longer-term aspiration.
The AI Leadership Capability Model (ALCM): A Self-Assessment
The AI Leadership Capability Model (ALCM) is WorkWise Academy's proprietary four-quadrant framework for assessing AI literacy at the senior leader level. It maps the four capability domains described above — Strategic Evaluation, Team Capability Judgment, Investment and Vendor Literacy, and Governance and Risk — and provides a self-assessment score for each.
Each quadrant is scored on a scale of 1-25, producing a maximum ALCM score of 100. The scoring is not based on knowledge recall — it is based on scenario judgment. For each quadrant, the assessment presents three to five realistic scenarios and asks the respondent to select the response that best reflects their current capability. Scores are calibrated against a validated rubric and should not be confused with the respondent's own confidence rating, which is consistently higher than the assessed score.
ALCM Score Interpretation:
- 20-40: Foundational. The leader has general awareness of AI but lacks the specific capabilities needed to make sound decisions across the four domains. AI investment decisions are likely to be based primarily on external advice and vendor presentations. Governance gaps are probable. Priority: structured AI literacy development, with particular focus on Strategic Evaluation and Governance.
- 41-60: Developing. The leader has working knowledge in one or two quadrants, typically Strategic Evaluation and/or Governance, but has significant gaps in Team Capability Judgment and Investment Literacy. AI decisions are inconsistent — sound in some areas, underdeveloped in others. Priority: targeted development in the lower-scoring quadrants.
- 61-80: Competent. The leader can operate effectively across all four quadrants for standard decisions. They have the literacy to evaluate proposals, assess vendor claims, read capability assessments, and confirm governance basics. Gaps are likely to appear at the edges of each quadrant — in complex or novel scenarios. Priority: scenario practice and peer learning to extend competence to edge cases.
- 81-100: Advanced. The leader can exercise nuanced judgment across all four domains, including complex scenarios. They are able to contribute substantively to AI strategy discussions at the board level and can mentor other senior leaders in developing their literacy. Priority: maintain currency as the AI landscape evolves, and extend capability into emerging domains (agentic AI, AI in regulated contexts, multimodal applications).
Most senior leaders completing the ALCM for the first time score in the 30-50 range. This is not a reflection of general capability — it is a reflection of the structured knowledge and practiced judgment that AI literacy specifically requires, which most senior leaders have not yet developed. An average ALCM score of 38 in a leadership team is typical for a well-run organisation that has not yet invested in executive AI literacy. An average of 71, which is achievable through a 90-day literacy sprint, represents a qualitatively different level of leadership capability on AI.
Case Study
A global management consulting firm with 120 partners across 4 offices ran a 90-day AI Literacy Sprint after identifying that AI strategy discussions were being dominated by 2-3 technically knowledgeable partners, with the majority unable to contribute meaningfully. The sprint included 3 structured briefings, 1 live build demonstration, individual ALCM assessments, and peer learning groups. Average ALCM score rose from 38 to 71 out of 100. The firm approved 4 AI investment proposals in the subsequent quarter that had previously stalled due to insufficient strategic consensus. See the full case study in Section 8.
Case Study: Global Consulting Firm, 120 Partners
A global management consulting firm with 120 partners across four offices in London, New York, Singapore, and Frankfurt identified in mid-2025 that its AI strategy was effectively stalled. The firm had made three significant AI-related investment proposals to the partner group in the preceding 12 months — a workforce capability development programme, an AI-assisted research platform, and an internal knowledge management system — all of which had generated inconclusive partner votes. The proposals were not rejected outright; they were simply unable to build sufficient consensus for approval.
The Managing Partner commissioned an internal diagnostic. The finding was consistent: AI-related agenda items at partner meetings were being dominated by two or three technically knowledgeable partners, whose arguments — however substantive — could not be evaluated, challenged, or built upon by the majority of the partner group. The result was deference rather than debate. Partners who could not assess the merits of an AI proposal defaulted to caution.
The Managing Partner decided to address this through a firm-wide AI Literacy Sprint, designed and delivered by WorkWise Academy, covering all 120 partners over 90 days. The programme was structured in three phases.
Phase 1 (Days 1-30): Foundation Briefings. Three structured briefings of 90 minutes each, delivered in person at each office location in rotating cohorts of 30. Content covered: what AI can and cannot do in a management consulting context (with specific examples from advisory, research, and delivery functions); how to evaluate an AI value proposition; and the five governance basics every partner should be able to confirm. The briefings were designed for busy partners — no pre-reading required, high information density, direct relevance to the firm's specific context.
Phase 2 (Days 31-60): Live Demonstration and Self-Assessment. A two-hour live build demonstration, attended by all partners in small groups of 15, in which a WorkWise Academy facilitator built a working market analysis tool from scratch using conversational AI, in real time, using a research question typical of the firm's client work. The demonstration was followed by a facilitated Q&A. All partners then completed the ALCM self-assessment and received individual scores and an action guide for their lowest-scoring quadrant.
Phase 3 (Days 61-90): Peer Application and Scenario Practice. Partners were assigned to peer learning groups of eight, crossing office boundaries, with the remit to apply AI to one real piece of client or firm work over the final 30 days and to share findings with their group. Three structured scenario exercises, distributed digitally and discussed in the peer groups, covered the three most common AI decision scenarios the firm faced: evaluating a vendor proposal, assessing team AI capability, and presenting a governance rationale to a client.
Results at Day 90:
- Average ALCM score across the 120 partners: 71/100 (up from 38/100 at baseline — an 87% average improvement)
- Lowest individual score at Day 90: 49/100; highest: 94/100
- 95 of 120 partners had used AI on at least one client or firm task during Phase 3
- Post-programme survey: 89% of partners said they could now engage substantively with AI-related agenda items at partner meetings; 91% said the live demonstration was the single most valuable element of the programme
In the quarter following the sprint, the partner group approved all four AI investment proposals that had previously stalled — including the three from the preceding year — with substantive debate about implementation design rather than division between proponents and uncertain abstainers. The Managing Partner described the shift as transformative: "We moved from a small number of AI advocates speaking into a room of politely uncertain partners, to a group of 120 people who could actually have the strategy conversation."
The programme also surfaced an unexpected benefit: seven partners identified as high-ALCM scorers became informal AI literacy champions within the firm, taking on a mentoring role with junior consultants and contributing to the firm's client-facing AI advisory practice. The literacy sprint, designed to improve governance and decision quality, created a secondary return through expertise diffusion.
Measuring Progress: The AI Literacy Index for Leadership Teams
A single ALCM assessment gives a point-in-time picture of a leadership team's AI literacy. Tracking progress over time requires a more lightweight measurement approach — one that can be administered quarterly without placing significant demands on executive time, while still producing reliable trend data.
The AI Literacy Index (ALI) is a condensed version of the ALCM, designed for quarterly administration. It consists of 12 scenario-based questions (three per quadrant) and takes approximately 15 minutes to complete. The ALI produces a score out of 60 that can be tracked against the baseline ALCM score and against team-level benchmarks from previous periods.
The primary value of the ALI is not the individual score — it is the team-level trend. Leadership teams that complete the ALI quarterly typically observe two patterns. First, progressive improvement in average scores, reflecting both the direct impact of any literacy development programmes and the incidental literacy gains from applying AI in practice. Second, periodic plateaus, which typically correspond to periods when the AI landscape has shifted (new capabilities have emerged, regulatory requirements have changed, the competitive context has evolved) and the team's existing knowledge has been partially obsoleted. Plateaus signal the need for a refreshed briefing, not a failure of development.
Three metrics are worth tracking at the team level:
- Average ALI score: the headline measure of team AI literacy, tracked quarterly. Target trajectory: 5-10 point improvement per quarter during active development, stabilising as the team reaches competent level.
- Quadrant distribution: the average score per quadrant across the team. Consistently low Governance scores (below 12 out of 25 on the ALCM equivalent) warrant specific attention — they indicate governance exposure that may not be visible at the aggregate level.
- Score range: the gap between the highest and lowest-scoring members of the leadership team. A wide range (over 30 points on the full ALCM) indicates that AI decisions are effectively made by a small subset of the leadership team, with the majority in a dependent or deferring role. Narrowing this range is a governance objective in its own right.
Getting Started: A 90-Day Literacy Sprint for Executives
The 90-day AI Literacy Sprint is the format that has consistently produced the most significant ALCM score improvements in the shortest time. It is structured in three phases corresponding to the three development approaches that work best for executives: foundation briefings, live demonstration, and applied scenario practice. The case study in Section 8 describes one implementation; the structure below is the reference model.
Phase 1: Foundation Briefings (Days 1-30)
Three structured briefings, each 90 minutes. Briefing 1: What AI actually is and does — calibrating capability expectations against production reality, not demo reality. Briefing 2: Where AI creates value in your industry and organisation — a value chain analysis tailored to the organisation's specific context, using real examples from comparable organisations. Briefing 3: Governance and risk — the five things that must be in place, and how to confirm they are.
The briefings should be delivered by a practitioner who also works with AI in the context relevant to the leadership team's industry, not a general-purpose AI educator. The credibility of the instructor and the specificity of the examples are critical to executive engagement. Generic AI content delivered to an executive audience without industry-specific relevance is perceived — correctly — as a poor investment of the leadership team's time.
Each briefing should include a 20-minute Q&A focused on the decisions the leadership team is currently facing: vendor proposals under review, capability investment decisions pending, governance questions that have been raised but not resolved. Connecting the briefing content directly to live decisions is what converts information into applicable literacy.
Phase 2: Live Demonstration and Self-Assessment (Days 31-60)
A single two-hour session combining a live build demonstration with individual ALCM assessment. The demonstration should build a tool or produce an output directly relevant to the leadership team's work — a strategic analysis, a competitive intelligence briefing, a financial model, depending on the organisation's context. The demonstration should show both AI's capability and its limitations: a prompt that works, a prompt that fails, and the facilitator's reasoning about why each outcome occurred.
The ALCM self-assessment is completed individually, privately, after the demonstration. Results are shared with the individual only (not aggregated for team comparison in this phase, to avoid social desirability effects in the initial assessment). Each leader receives an individual action guide for their two lowest-scoring quadrants.
The live demonstration is the element of executive AI literacy development that research and practitioner experience most consistently identify as highest impact. The reason is straightforward: watching AI work in real time — and work incorrectly, and be corrected — creates an accurate capability model that no amount of description or explanation can replicate. A 30-minute live demonstration creates more durable and accurate literacy than a 3-hour lecture on the same subject.
Phase 3: Peer Application and Scenario Practice (Days 61-90)
The final phase converts literacy into applied judgment through two mechanisms. First, each leader applies AI to at least one real decision or work task during the 30-day period, using the action guide from their ALCM assessment to focus on their development areas. Second, three structured scenario exercises are completed in peer groups of 4-6, each presenting a realistic AI decision scenario and requiring the group to reach a reasoned conclusion.
The scenarios should be calibrated to the decisions the leadership team is actually facing: a vendor evaluation, a capability investment decision, a governance question, a client disclosure challenge. The quality of the peer discussion — not the individual answer — is the development mechanism. Leaders who have done the foundation work in Phase 1 and the demonstration in Phase 2 are now testing their literacy against each other's, in a low-stakes context where reasoning can be challenged and refined.
At Day 90, administer the ALI to establish the post-sprint baseline. Compare to the Phase 2 ALCM scores. Set the agenda for Phase 4 — which, for most leadership teams, is a regular quarterly briefing and annual ALCM refresh to maintain currency as the AI landscape continues to evolve.
Key Takeaways
- Only 12% of C-suite executives can confidently evaluate AI-generated outputs without assistance; 81% say AI is a top-three strategic priority. The gap between stated priority and actual capability is the primary strategic risk in executive AI governance.
- The four capabilities of AI-literate senior leaders are Strategic Evaluation (knowing where AI creates real value), Team Capability Judgment (accurately assessing what their people can do), Investment and Vendor Literacy (evaluating proposals without being misled), and Governance and Risk (knowing the five things that must be in place).
- The AI Leadership Capability Model (ALCM) provides a self-assessment score across all four quadrants, on a scale of 0-100. Most leadership teams score 30-50 on first assessment; a 90-day literacy sprint routinely produces improvements to the 65-75 range.
- A 30-minute live AI build demonstration creates more durable and accurate literacy than a 3-hour lecture. The mechanism is direct experience of AI's actual capability and limitations — which cannot be accurately conveyed through description.
- Investment literacy means being able to ask the 10 questions that reveal whether an AI vendor has a tested, production-ready product or a compelling demo. The quality and specificity of a vendor's answers to these questions are the primary signal of product credibility.
- Governance literacy means being able to confirm the five governance basics are in place: a data classification policy for AI inputs, a quality control standard for AI outputs, an AI disclosure framework, a vendor due diligence protocol, and a clear accountability structure.
- The 90-day AI Literacy Sprint structure: Phase 1 (Days 1-30) three foundation briefings; Phase 2 (Days 31-60) live demonstration plus ALCM self-assessment; Phase 3 (Days 61-90) peer application and scenario practice. Maintain currency with quarterly ALI benchmarks thereafter.