Harvey AI's 70% DAU/MAU and Enterprise AI Deployment ROI

Read by leaders before markets open.
Allen & Overy's deployment of Harvey AI in late 2022 produced a finding that has since reframed how legal departments justify AI spending: daily active user rates above 70% proved more persuasive to CFOs than any cost-per-document model. Harvey reached a reported $3 billion valuation in 2025, and the engagement data from early enterprise clients drove that outcome as much as any product feature.
The signal was not contract volume processed. It was DAU/MAU ratios consistently above 70%, a benchmark more commonly associated with consumer social apps than B2B legal software, according to SaaStr's analysis of Harvey's go-to-market model.
What This Case Study Actually Tested
This analysis draws on Harvey's publicly disclosed deployment data, SaaStr's coverage of Harvey's B2B engagement metrics framework, and published outcomes from Allen & Overy's Harvey rollout beginning in late 2022. No primary interviews informed this piece. Where numbers are attributed to Harvey directly, the source is Harvey's investor communications or public press releases.
The core question Harvey's data answers is narrow but important: can daily engagement rates serve as a reliable proxy for enterprise AI ROI in knowledge-work deployments? The sample covers legal services firms, primarily large international law firms and Big Four professional services practices. Manufacturing, logistics, and financial services deployments are outside the scope.
The timeframe covers Q4 2022 through Q1 2026. Harvey's deployment model is a vertical-specific large language model built on foundational models from Anthropic and OpenAI, fine-tuned on legal corpora and integrated into document workflows attorneys already use.
Does Enterprise AI Deployment ROI Depend on Engagement Metrics Like DAU/MAU?
Harvey's DAU/MAU data above 70% proves workflow integration, not work-product quality. When lawyers open a tool on 21 out of 30 working days, the tool has become part of daily practice, signaling budget defensibility in ways a cost-per-transaction model cannot. At Allen & Overy, this pattern translated directly to a $3 billion valuation narrative that resonated with CFOs faster than any per-document calculation.
Allen & Overy, which employs roughly 4,000 lawyers globally, deployed Harvey in late 2022 across selected practice groups. Lawyers using Harvey reduced contract-review time by approximately 50% on standard commercial agreements, according to Harvey's published case materials. PwC followed with a deployment across its legal services division, citing similar efficiency gains in due diligence document review.
Typical Enterprise SaaS vs Harvey AI: 90-Day Active User Retention
Harvey reports that its top enterprise clients show DAU/WAU ratios above 85%, meaning lawyers who use Harvey on any given week return the following day with high frequency. That pattern differs from the typical enterprise SaaS adoption curve, where tools purchased in Q1 see active use drop by 60-70% within 90 days, according to Gartner's enterprise software adoption research.
For legal department heads justifying a $500,000-plus annual contract to a CFO, a DAU chart that resembles a consumer app is a more persuasive budget artifact than a cost-per-document model built on contested assumptions about billable hour rates.
Harvey AI: Contract Review Time Reduction by Task Type
The 70% reduction in legal research summarization time carries the highest per-hour value in Harvey's published data. Legal research is a senior associate task billed at $400 to $800 per hour at major firms. If Harvey compresses a four-hour research task into 70 minutes, the per-engagement economics become straightforward before engagement rates even enter the conversation.
KEY TAKEAWAY: Harvey's central insight is that DAU/MAU above 70% is a stronger ROI signal than cost-per-transaction in legal AI, because it proves workflow integration rather than occasional use. A tool lawyers open every day is not shelfware. A tool that reduces cost-per-document by 40% but sits idle three weeks per month is.
Why Enterprise Buyers Misread Engagement Data
Three misuse patterns appear regularly in legal AI procurement conversations.
First, buyers conflate engagement with accuracy. High DAU/MAU means lawyers open Harvey daily. It does not mean every output Harvey produces is correct. Harvey's own documentation states that attorney review remains mandatory for all substantive outputs. Firms that allow DAU rates to crowd out accuracy audits from their quality-assurance process create malpractice exposure.
Second, buyers apply peer-firm DAU benchmarks without adjusting for practice area. A litigation support team using Harvey for deposition summary drafts will show different engagement patterns than a transactional team using it for purchase agreement markup. Applying Allen & Overy's headline DAU numbers to a specialist boutique with a different work mix produces a meaningless comparison.
Third, CFOs sometimes accept DAU as a stand-alone justification without demanding conversion metrics: did higher engagement translate into more matters handled per attorney, lower associate overtime, or measurable revenue per lawyer? Engagement metrics answer whether the tool is used. They do not answer whether the business improved without a second data layer.
What Harvey's Deployment Record Does Not Prove
This deployment record does not establish five things buyers commonly assume.
First, Harvey's DAU metrics do not prove the engagement framework transfers to non-legal knowledge work. The legal domain has unusually clear task structures (document review, research, drafting) that make AI integration easier to measure. HR, strategy, or marketing functions have less defined workflows.
Second, the 50% to 70% time reduction figures do not account for rework. If Harvey outputs require substantial attorney correction, the gross time saving shrinks. Harvey's published materials do not disclose rework rates or error correction time.
Third, high engagement does not prove profitability. A firm paying $3,000 per attorney seat annually and achieving 50% faster contract review still needs to determine whether that time saving generated additional revenue or simply reduced billable hours.
Fourth, Harvey's enterprise results do not generalize to legal departments with fewer than 20 attorneys. For smaller teams, the fixed cost of implementation, training, and IT integration represents a larger share of any potential saving.
Fifth, Harvey's valuation trajectory from seed to $3 billion reflects investor expectations about legal AI's total addressable market, not a validated proof that Harvey's ROI model holds at scale across diverse firm types.
How Should CFOs Use Agentic AI Workflow Automation Metrics to Govern Legal AI Spend?
CFOs who treat DAU/MAU as a single-layer accountability metric miss a structural gap: engagement proves adoption, not business value. The most defensible governance model pairs DAU/MAU thresholds with a second operational metric, specifically revenue per attorney or matters handled per attorney, measured at 90-day intervals from deployment. For a $500,000-plus legal AI contract, both numbers belong in every budget review cycle.
Harvey's deployment stall data reinforces this framework. Partner resistance accounts for 21% of legal AI deployment stalls, according to Gartner's enterprise AI adoption survey. High platform DAU can mask near-zero influence on actual work product if senior partners reject AI-drafted outputs at the review stage. CFOs who build dual-metric accountability into vendor contracts catch this failure mode before the renewal decision.
Primary Reasons Enterprise Legal AI Deployments Stall at 90 Days
Four additional friction points emerge consistently in enterprise legal AI rollouts.
Data security conflicts arise when firm IT policies prohibit uploading client documents to third-party systems. Harvey operates under enterprise data agreements, but firms with sovereign data requirements, particularly those serving government clients or operating under EU data residency rules, face contractual barriers no engagement metric resolves.
Model versioning creates audit risk. When Harvey's underlying model updates, outputs for identical inputs can change. Law firms with document retention obligations need to know which model version produced which output. Harvey provides version controls, but smaller legal departments often lack the IT staffing to implement them correctly.
Integration depth determines actual time savings. Firms that deploy Harvey as a standalone tool without connecting it to their document management system (iManage or NetDocuments) see materially lower DAU rates. The full 50% to 70% time reduction requires Harvey to sit inside the document workflow.
Billing model misalignment creates perverse incentives at firms still billing by the hour. If Harvey saves an associate four hours on a research task, the firm may bill the client four fewer hours. Senior partners see revenue compression before they see efficiency gains. This structural tension is the primary reason large law firms move slowly on legal AI despite high associate enthusiasm.
Enterprise Legal AI Market Size Projection ($B)
The legal AI market is projected to grow from $900 million in 2023 to $9.2 billion in 2026, according to Goldman Sachs legal tech research. Firms that lock into multi-year enterprise contracts before establishing DAU-based evaluation checkpoints face significant switching costs if engagement drops.
What This Means for COOs, Legal Heads, and CFOs
For COOs managing legal department budgets, Harvey's DAU framework provides a genuine reporting advance over prior legal technology ROI models. Build a simple checkpoint into every legal AI contract: require DAU/MAU data at 90 days and 180 days. Tools below 50% DAU/MAU at six months rarely recover. For a related framework on how engagement metrics translate into enterprise AI budget governance, see our enterprise AI ROI analysis covering four practices that unlock 55% returns.
For legal department heads evaluating Harvey against competitors (Casetext, Thomson Reuters CoCounsel, LexisNexis AI), Harvey's published DAU benchmark creates a comparison point rivals have not yet matched publicly. Require every vendor to provide DAU/MAU data from comparable-size deployments, not just outcome case studies. The vendor that refuses this disclosure likely has low engagement numbers.
For CFOs approving legal AI budgets above $500,000 annually, demand the second data layer engagement metrics alone cannot provide: revenue per attorney or matters handled per attorney, measured before and after deployment. DAU proves adoption. Revenue-per-attorney proves value. Both numbers belong in the budget presentation. Our agentic AI workflow automation CFO frameworks analysis covers the governance structure for requiring dual-metric accountability from AI vendors.
Legal departments expanding Harvey to client-facing work product should also review the compliance considerations engagement data does not address. AI-generated legal work falls into an ambiguous zone under attorney professional responsibility rules in multiple jurisdictions. The agentic AI regulatory gap analysis covers adjacent regulatory risk.
Limitations of This Analysis
This analysis relies entirely on vendor-published data and third-party market research. Harvey has not released audited outcome data, independent accuracy benchmarks, or rework rates. The Allen & Overy and PwC figures come from Harvey's own case materials, which select for positive outcomes. No independent academic study has replicated Harvey's reported DAU or time-reduction figures across a representative sample of firms. Readers should treat the specific percentages as directional rather than precise. The market projection figures from Goldman Sachs reflect analyst estimates subject to the standard range of error in early-stage market sizing.
Clear Verdict
Harvey's engagement-first ROI framework is a genuine improvement over cost-per-transaction models in legal AI procurement. Build DAU/MAU checkpoints into every AI vendor contract now, before the market normalizes around it and vendors stop treating the data as a differentiator.
The critical caveat buyers miss: engagement is measurable immediately, while revenue impact takes 12 to 18 months to appear. Vendors know this and use DAU data precisely because it arrives before profitability evidence does. If your firm's DAU/MAU falls below 50% at 90 days, trigger the remediation clause immediately. If DAU does not recover to 60% within 60 days of remediation, the tool has not achieved workflow integration and the budget should be redirected.
For legal departments with hourly billing models, the ROI calculus may not close regardless of engagement rates. Measure Harvey's value through capacity expansion instead: can the same team handle 20% more matters at the same quality standard? For fixed-fee or subscription-billing departments, Harvey's published numbers support a credible cost-reduction business case directly.
Sources
- SaaStr, "DAU, WAU, and MAU Are the New Lighthouse Metric in B2B AI: Harvey's a Great Case Study." saastr.com
- Gartner, "Enterprise AI Adoption Survey 2025." Publication cited by name; no verified URL available.
- Goldman Sachs, "Legal Tech Research 2025." Publication cited by name; no verified URL available.
- Harvey AI, "Allen & Overy and PwC Deployment Case Materials." Publication cited by name; no verified URL available.
Frequently Asked Questions

Red Hat's 233% ROI: enterprise AI deployment proof points
Forrester Consulting validated 233% ROI and 6-month payback for enterprise AI deployment on Red Hat OpenShift AI. Learn which conditions apply to your organization.

$10B DeployCo: Agentic AI Governance Framework Gaps
OpenAI's $10B DeployCo embeds engineers inside PE portfolio firms, bypassing IT procurement. CFOs need an agentic AI governance framework before deployment starts.

Microsoft MAI Models Reshape AI Procurement Strategy 2026
Microsoft's MAI models launch April 2026 at $0.36/hr, splitting Azure and OpenAI pricing. Here's what CTOs and CFOs must do now to renegotiate AI contracts.