ENGRAM
Thesis Technology Team Get in Touch

Superhuman Data
for Finance

The reasoning layer for long-horizon finance agents.

View the Thesis Schedule a Conversation
scroll
HOW IT WORKS

Three steps. One session.
Training-ready data.

1
Expert records a session
An analyst works through a real financial problem — reasoning aloud while their screen is captured.
2
Engram decomposes the reasoning
Cognitive events are detected, timestamped, and cross-referenced against screen behavior — automatically.
3
Your models learn how experts think
Training-ready JSONL with temporal signals, screen context, and multi-scale reasoning chains.
Expert reviewing CIM cover page
00:23:41 CONCERN conf 0.94
"Wait — the EBITDA margin assumption doesn't hold if you back out the one-time items. Let me go back to the source filing..."
pause: 4.2s · revision_triggered: true
Expert reviewing financial tables
00:24:03 INVESTIGATION conf 0.89
"OK so looking at the 10-K, note 14... yeah, there's a $12M restructuring charge they're adding back. That's aggressive."
pause: 0.8s · evidence_type: "regulatory_filing"
Expert reviewing financial projections
00:25:18 CONCLUSION conf 0.97
"So the real margin is closer to 18%, not 24%. That changes the entry multiple meaningfully — I'd want to see 6x, not 8x."
judgment_revised: true · delta: -25% margin
1 reasoning arc · 3 events · 97s elapsed · confidence 0.94 → 0.97
THE PROBLEM

Model performance drops exponentially on multi-step tasks

Scaling laws have driven AI performance through more compute, larger models, and more tokens. The binding constraint is now data quality. In verifiable domains — coding, mathematics — models can self-correct through automated feedback; synthetic data and training environments suffice. In open-ended domains like finance, law, and medicine, correctness depends on contextual judgment with no ground-truth signal to train against. These domains require a fundamentally different data infrastructure — and it does not yet exist.

Current models are trained on outcomes — what an expert concluded, structured after the fact. The reasoning moves that drive per-step accuracy — the cross-reference that triggered a pivot, the assumption revised midway through a discounted cash flow, the confidence shift before a final judgment — are absent from every existing training corpus. The data that exists captures decisions. Engram builds data that captures the process of deciding.

P(Success) = AccuracySteps
99% per-step accuracy 99.9% per-step accuracy
~$300B
AI software market by 2027, shifting to agentic execution where per-step accuracy determines commercial viability
Gartner Forecast: AI Software 2023–2027
84%
of PE firms have appointed a Chief AI Officer, with reliability on complex multi-step tasks as the binding constraint
EY US Private Equity AI Insights, 2025
<1% → 33%
of enterprise software will include agentic AI by 2028, up from less than 1% in 2024 — training data quality is the primary differentiator
Gartner Newsroom, Aug 2025
THE DATA

Reasoning traces captured in real time, structured for model training.

Every competitor asks experts to write down their reasoning after the fact. The result is a reconstruction — shaped by hindsight, narrative instinct, and the impulse to sound coherent. Engram captures reasoning while it happens. Here's the difference.

unit_event.json — single cognitive event
{
  "event_type": "CONCERN",
  "start_ts": 1421.3,
  "end_ts": 1438.7,
  "transcript": "Wait — the EBITDA margin assumption doesn't hold if you back out the one-time items...",
  "confidence": 0.94,
  "pause_delay": 4.2,
  "screen_state": "event_1421.300.jpg",
  "revision_triggered": true,
  "label": "EBITDA margin challenge"
}
event_type
Cognitive event classified in real time — not a retrospective label. The expert didn't write "CONCERN." Their behavior revealed it.
pause_delay: 4.2s
A 4-second pause before speaking. In retrospective annotation, this signal — uncertainty forming before articulation — is invisible.
screen_state
The exact screen the expert was looking at. Voice cross-referenced against visual context. The data is self-verifying.
revision_triggered
This event caused the expert to revise a prior conclusion. No annotation platform captures when and why a judgment changed.
THE TECHNOLOGY

Codifying Wisdom into
Reasoning Trace Packages

Deterministic Multimodal Reproducible
Local
Expert Session
Voice + screen captured concurrently
›
Event detection Frame extraction Temporal enrichment
1 API Call
Cognitive Event Mapping
Structured event data with temporal signals
Local
Human Dashboards
Explainability & session analytics
Local
SFT Training Bundle
Multi-scale, multimodal JSONL
One LLM call per session. Everything else is local and deterministic.
Why Engram Data Drives Better Model Performance
Concurrent capture with temporal structure

Existing training data asks experts to reconstruct their reasoning after the fact — producing accounts contaminated by rationalization and hindsight. Engram captures reasoning as it happens, preserving the signals that retrospective annotation destroys: the 4-second pause before an insight, the confidence shift after cross-referencing a source, the pace change that preceded a revision. Time is information. The temporal architecture of expert reasoning is a first-class training signal, not metadata.

Multi-scale training with embedded preference signal

Each session produces three interlocking training signals: event classification (per-event cognitive typing), reasoning chains (multi-step logic reconstruction), and session synthesis (full-session analysis). The deliberative structure — hesitation, revision, hypothesis rejection — is encoded directly in the trace. Standard SFT requires extensive RLHF to correct errors the training data failed to prevent. When the preference signal is already in the data, RLHF becomes a refinement step rather than a corrective one.

Multimodal grounding, not text-only

Every training example embeds the screenshot the expert was looking at when they spoke — base64-encoded inline via VISION format. The model learns situated reasoning: cognition grounded against the specific financial data on screen. Voice cross-references visual context, making the data self-verifying by design.

Compounding domain schema

Each session refines the reasoning format itself. As the corpus grows, it accumulates domain-specific structure — branching patterns, edge-case heuristics, decision-point taxonomies — that later sessions build upon. Session 200 is structurally richer than session 10. The result is a corpus that compounds in depth, not just volume, and grows harder to replicate with every session.

HOW EXPERTS EARN

Your Knowledge Compounds.
So Do Your Royalties.

The prevailing compensation architecture for AI training data decouples expert contribution from downstream value creation. Annotators are paid per task or per hour — a transactional model that extinguishes the contributor's economic relationship with their knowledge at the point of delivery. The resulting asymmetry is structural: the expert's judgment compounds indefinitely inside the models it trains, while the expert captures none of that compounding value.

Incumbent model
$150/hr
One-time extraction.
No residual claim.
Compensation is indexed to time, not to the epistemic value of the contribution. A senior credit analyst's twenty years of pattern recognition is priced identically to a first-year associate's — by the hour.
Engram model
Royalty
Per trace, per customer.
Income compounds with demand.
Compensation is indexed to downstream usage. Each time a reasoning trace is selected for a training run, the contributing expert receives a royalty. The income stream scales with demand — aligning the expert's financial incentives with the long-term value of their intellectual contribution.

The heuristics a portfolio manager has refined across multiple market cycles, the judgment a CFO has built through a hundred transactions — this is an intellectual legacy. We are building the infrastructure for humanity's epistemic endowment: a system in which the knowledge of its most capable practitioners is preserved in structured form, and returned to them — with every model that learns from it.

WHO IT'S FOR

Cognitive infrastructure for AI companies building in finance.

Any team whose competitive advantage depends on the quality of financial reasoning their models can produce.

01
Financial intelligence platforms
Vertical AI companies whose product is financial reasoning — document analysis, credit assessment, due diligence automation. Their model quality is their product quality, and that starts with training data that encodes how experts actually think, not what they wrote after the fact.
02
PE firms and hedge funds
Want AI associates that evaluate CIMs the way their best partners do. The moat isn't the model architecture — it's whose judgment it encodes. A proprietary model trained on how your senior deal team actually reasons is a fundamentally different asset than a generic financial LLM.
03
Strategy and advisory firms
When a senior partner leaves, their approach to evaluating a platform company leaves with them. Training AI on that reasoning means the firm's intellectual capital compounds instead of depreciates. The institutional knowledge persists beyond any individual career.
TEAM

Built by Operators from
Finance and AI

Advised by operators who have scaled data products, led enterprise transactions, and deployed capital at the frontier of finance and technology.

Avantika Mehra
Co-Founder & CEO
Avantika Mehra
UVA Cognitive Science Bain PEG UVA Memory Lab Cosmos Institute
  • EEG research at UVA Computational Memory Lab — extracting latent heuristics from expert behavior
  • 30+ analyses at a growth equity firm via Bain's Private Equity Group
  • Chief of Staff at Cosmos Institute
Raghav Singh
Co-Founder & CTO
Raghav Singh
UC Berkeley EECS McKinsey QuantumBlack BAIR Lab
  • Senior Data Engineer at QuantumBlack (McKinsey AI) — 5 Fortune 100 agentic AI deployments
  • 2.5 years deploying AI solutions for institutional finance and strategy clients
  • Explainability research at Berkeley AI Research Lab (BAIR)
WHERE WE'VE WORKED
Bain & Company McKinsey QuantumBlack Cosmos Institute Avataar AI Mastercard
STRATEGIC ADVISORS
John Doe
Associate Partner, Bain & Company
  • 15+ years in PE due diligence and portfolio operations
  • First expert to contribute reasoning traces to Engram
Raoul Nanavati
CEO, Voice AI Startup (Financial Services)
  • Built voice AI infrastructure for banking and insurance
  • AI product strategy for regulated industries
Gerald Doe
DPhil Researcher, Oxford Internet Institute
  • Stanford Philosophy (Suppes Award) + CS
  • Published at ACL, EMNLP, ICML, NeurIPS
  • AI evaluation methodology and trust
Jay Doe
Partner, Bain & Company
  • Head of Global Innovation & Design practice
  • 25+ years in consulting — UVA Engineering + Darden MBA
  • PE relationships across India, Middle East, and US
Jane Doe
Graduate Researcher, MIT Media Lab
  • Fluid Interfaces group — human-computer interaction
  • Cognitive interface design and reasoning capture
MULTI-MODAL
Voice + screen captured concurrently
SESSION-COHERENT
Complete reasoning arcs, not fragments
SECURITY-FIRST
Encrypted in transit. Role-based access. Audit-logged exports.
WHAT PEOPLE ARE SAYING
"This is the first dataset I've seen that captures how analysts actually think — not what they say they think after the fact."
Partner, Top-3 Strategy Firm
"We've tried every annotation platform. None of them produce data with temporal structure. Engram's traces are fundamentally different."
Head of AI, Finance-Vertical AI Company
GET IN TOUCH

Interested in Engram?

We're working with a select group of design partners. Schedule a conversation to learn more.

Schedule a Conversation
ENGRAM · New York, NY · 2026 · Privacy · Terms