Reproducibility in GEO: Deadline Passed, Verification Begins
Date: Tuesday, November 4, 2025
On Monday, November 3, 2025, the submission deadline for GEO vendor reproducibility tests closed. No submissions were received.
This does not signal failure of any individual company. It signals the current maturity of the category. GEO products are evolving fast, yet independent verification workflows are not yet established. Visibility claims remain ahead of audit-grade evidence.
Summary of requested protocol
Vendors were asked to demonstrate reproducibility within a controlled window:
• 24 prompts across two assistants and two regions
• 3 runs per prompt per assistant within 48 hours
• Reproducibility tolerance: inclusion within ±5 percentage points and average rank within ±0.5
• Required artifacts: UTC-timestamped run logs, model or build identifiers where exposed, prompt text and parameters, and SHA-256 evidence hashes for each run
This protocol was designed to be practical, limited in scope, and aligned with audit expectations when outputs influence strategic decisions or external narratives.
Why this matters
AI assistants shape brand perception, investor sentiment, and competitive context in real time. Once these signals influence planning, disclosure language, or board-level understanding, independent evidence becomes a control requirement. Reproducibility is not a preference for transparency. It is the basis for governance.
Phase two now begins
• AIVO will run independent reproducibility and variance tests across leading assistants
• All runs will be time-stamped, hashed, and recorded in a reproducibility ledger
• Evidence packs will be prepared for audit, risk, finance, and communications teams
• Initial briefing set for Friday, November 14, 2025, followed by a public summary the next week
Late submissions
Vendor submissions remain welcome. Any received after November 3 will be recorded as post-deadline and assessed against the same thresholds.
Anticipated questions
Frequent model updates
Model evolution is expected. Reproducibility is evaluated within fixed windows and logged with model or build identifiers where disclosed. Out-of-tolerance results are treated as a change signature, not a failure.
Data sensitivity
Evidence can be anonymized. Verification concerns the existence of reproducible outputs, not disclosure of proprietary models or clients.
The signal
Markets mature when claims coexist with proof. GEO is entering that phase. Verification will move alongside innovation.
Evidence, not assertion, becomes the basis of trust.
Access to materials
• Protocol Pack v1.2: available on request
• Logging Framework: available on request
• Submission Guidelines: available on request
• Reproducibility Ledger Q4-2025: available on request
• Secure Submission Uploader: available on request