Tools
3 — calculator · mapper · glossary
License
CC-BY 4.0 · embed allowed
Data
Calibrated to 44 cases
Updates
Quarterly
Resources

Three small tools that make the framework operational.

The assessment tells you where you stand. These tools tell you what to do about it: a payback calculator that uses the case-derived diffusion curve, a task-mapper that scores six common workflows on RAPID-fit, and a glossary so the language stops being a barrier.

§ 01

Payback calculator.

Plug in seats, salary, time-on-task lift. The diffusion curve from the case dataset bends the result toward reality.

$
%
%
$
×
Realistic annual net benefit
$5.5M/yr

Payback in 38 days. If you raised active-user rate from 42% → 70% (the Klarna benchmark), net benefit climbs to $9.2M/yr. The framework's case data shows that lift is achievable inside one quarter with a named change-management lead per BU.

Open full calculator
§ 02

Task-RAPID mapper.

For six everyday workflows, a per-dimension fit score derived from the case dataset. Use it before committing budget.

Task
Readiness · data fit
Alignment · strategy fit
Impact · measurable lift
Diffusion · adoption ease
L1 customer supportKlarna · DPD
82High
76High
88High
61Medium
Code generation / dev assistCopilot · Cursor
88High
64Medium
79High
84High
Knowledge worker assistantMorgan Stanley AAA
74High
81High
58Medium
77High
Marketing copy generationCNET · Bankrate
58Medium
42Low
38Low
72High
Clinical decision supportMayo · Epic
44Low
72High
76High
38Low
Financial forecastingZillow · iBuying
36Low
52Medium
69High
31Low
Open full task mapper
§ 03

Glossary.

30 terms used across the assessment, report and case dataset — defined the way RAPID uses them.

4
44-case datasetDataset

The curated, primary-source-linked corpus behind the framework. Each case has an organisation, period, industry, function, outcome, evidence statement, and source URL.

Internal
A
Alignment (A)Framework

Strategic alignment: how well GenAI investments are tied to measurable business objectives with named executive sponsorship.

Technology Acceptance Model (TAM)
Adoption AuditFramework

A five-element checklist that predicts deployment success: business objectives, performance tracking, change management, iterative deployment, and a named executive sponsor.

Internal
Anchor sourceMethodology

A specific cited case or academic citation that supports a claim in the framework. The Task Mapper labels every quantitative claim either anchored or illustrative.

Internal
B
Blind spotDataset

On the team-comparison view (/compare/team): a dimension where every participant scored below 40%. Likely the cheapest place for the team to improve together.

Internal
C
Case IDMethodology

Stable identifier (CASE-001 through CASE-045) for each entry in the 44-case dataset. Used in citations and cross-references throughout the site.

Internal
Confidence intervalDataset

Width of the band around an industry benchmark, reflecting sample size. Industries with fewer than 10 cases (Healthcare, Manufacturing, Government) carry wider intervals.

Internal
D
Diffusion (D)Framework

Organisational adoption: change management, user training, and cross-functional collaboration that turn a deployment into a behaviour change.

Diffusion of Innovations
DeLone & McLean IS SuccessTheory

DeLone & McLean (2003). Six interdependent dimensions of IS success: information quality, system quality, service quality, use, user satisfaction, and net benefits.

DeLone & McLean IS Success Model
Diffusion of InnovationsTheory

Rogers (2003). Innovations spread through social systems via adoption categories: innovators, early adopters, early majority, late majority, laggards.

Diffusion of Innovations
Design Science Research MethodologyMethodology

Peffers et al. (2007). The development methodology behind RAPID: identify problem, define objectives, design artefact, demonstrate, evaluate, communicate.

Peffers 2007
Deployment stageMetric

Where a case sits on the maturity arc: Pilot/POC, Scaling, Production, Optimization, or Abandoned. The 44-case dataset is weighted toward Production and Optimization.

Internal
F
Failure modeMethodology

One of six categories where GenAI deployments tend to break down: data, technical, organisational, strategic, measurement, economic.

Internal
I
Impact (I)Framework

Measurement maturity: KPI definition, baseline establishment, and ROI review processes for GenAI initiatives.

DeLone & McLean IS Success Model
IT Productivity ParadoxTheory

Brynjolfsson & Hitt (1993). IT productivity gains require organisational change with a 3-5 year lag. AI productivity figures show the same pattern - per-task minute savings do not aggregate automatically into firm-level productivity.

Brynjolfsson 1993
K
Kotter 8-Step Change ModelTheory

Kotter (1996). Eight sequential steps for organisational change: urgency, coalition, vision, communication, empowerment, wins, consolidation, anchoring.

8-Step Change Model
M
Maturity LevelFramework

Buckets that summarise an overall RAPID score: Nascent (0-25%), Developing (26-50%), Established (51-75%), Advanced (76-100%).

Internal
P
Portfolio (P)Framework

Portfolio balance: diversification across risk levels and time horizons; the discipline to kill or scale initiatives based on results.

Portfolio Selection Theory
PercentileMetric

Where an organisation's score sits relative to peers in the dataset. The 60th percentile means 60% of peers scored at or below this point.

Internal
POC abandonmentMetric

Share of GenAI proofs-of-concept that never reach production. Gartner forecasts 30% by end of 2025; BCG finds only 22% of organisations advance past POC at all.

Internal
R
RAPIDFramework

A five-dimension maturity framework for GenAI investment - Readiness, Alignment, Portfolio, Impact, Diffusion - that scores an organisation against a 44-case enterprise dataset.

Internal
Readiness (R)Framework

Data and technical readiness: data infrastructure, MLOps maturity, model-evaluation discipline, and AI/ML talent depth.

UTAUT (Unified Theory of Acceptance)
ROI rangeMetric

Industry-specific projected return on GenAI investment, derived from thesis Chapter 4. Tech 18-24%, Financial Services 15-22%, Healthcare 12-18% are the most-studied bands.

Internal
S
Success rateMetric

Share of measurable cases (success + failure + partial) in a cohort that achieved their stated objective. Excludes Too Early outcomes.

Internal
T
Technology Acceptance Model (TAM)Theory

Davis (1989). Predicts technology adoption from perceived usefulness and perceived ease of use. Strongest predictor of whether a tool gets used at all.

Technology Acceptance Model (TAM)
Tier 1 sourceMethodology

Peer-reviewed journals, NBER working papers, government findings, and investigative-journalism cases with documented evidence trails.

Internal
Tier 2 sourceMethodology

Business press, corporate disclosures, and analyst reports with primary-source URLs but lower formal-vetting depth than Tier 1.

Internal
Time to ROIMetric

Typical payback window for GenAI investments by industry. Ranges from 12-18 months (tech, retail) to 24-36 months (government, regulated sectors).

Internal
U
UTAUTTheory

Venkatesh et al. (2003). Unifies eight technology-adoption models. Performance expectancy, effort expectancy, social influence, and facilitating conditions explain ~70% of adoption variance.

UTAUT (Unified Theory of Acceptance)
Use-case typeMethodology

One of six GenAI deployment archetypes: Automation, Generation, Analysis, Augmentation, Chatbots, Decision Support. Each has a verified success-rate band derived from the dataset.

Internal
Open full glossary

Use these in your own deck. Embed any of them with one line.

Each tool is embeddable with a snippet — preserves attribution and updates automatically as v2.0 ships patches. CC-BY 4.0.