OMES seeks to accelerate AI adoption across Oklahoma's 125 state agencies by deploying self-contained "engineering pods" that deliver production-ready AI capabilities. Rather than staff augmentation, the state wants outcome-based partnerships that leave behind working software and upskilled internal teams.
We have proposed three potential starting points for an initial test of this AI pod model: (1) an AI-powered legislative research assistant, (2) a statewide chatbot framework, and (3) a SNAP QA system designed to begin the process of decreasing the payment error rate. We have also identified a number of Tier 2 projects that could be of interest. Our immediate next step is to work with you to determine prioritization and scoping for these projects - which are top priorities and achievable over a relatively short time horizon with a modest resource investment.
Key Themes from Discussion
Move from pilot to production - no more proof-of-concepts, need real deployables
Pod-based delivery model (not headcount scaling)
Knowledge transfer embedded - OMES staff work alongside Concourse engineers
Outcome-based pricing preferred over hourly billing
Technology-agnostic across clouds (Azure, GCP, AWS) and models (OpenAI, Anthropic, Gemini)
Working Demos
Important Caveat: These demos are conceptual prototypes built before any discovery conversation. They are intended to illustrate technical capabilities and spark discussion - not to represent final solutions. Actual implementation would follow proper discovery, requirements gathering, and stakeholder input.
A natural language Q&A system for Oklahoma legislative staff to research statutes, bills, and legal documents.
Note: This demo uses publicly available sources; production will use authoritative sources with proper governance.
Demo for Discussion
Data Pipeline: Downloaded all 86 Oklahoma statute titles as PDFs, extracted text, chunked into searchable segments
Search Index: Indexed all statute chunks in TurboPuffer with hybrid search (BM25 + vector) - 87k+ searchable chunks
LegiScan Integration: Connected to LegiScan API for bill data (28 sessions, 2010-2026, 5,904 current session bills)
Chat Interface: Next.js web app with natural language Q&A, streaming responses, and citation display
Model Router: Abstraction layer supporting OpenAI, Anthropic, and Google models
Current Architecture
Architectural Decisions
Why a Vector Database? Vector databases enable semantic search - finding documents by meaning rather than just keywords. When a user asks "What are the rules for charter schools?", the system finds relevant statutes even if they don't contain those exact words.
Why TurboPuffer? We have evaluated several vector database options including Pinecone, Qdrant, Weaviate, and Chroma. We selected TurboPuffer for its excellent query performance, hybrid search capabilities (combining vector and keyword search), and competitive pricing. The architecture is portable - state-approved alternatives (Pinecone, PostgreSQL pgvector, Azure AI Search, etc.) can be substituted based on procurement requirements.
Data Sources: Additional data sources (Oklahoma Constitution, AG opinions, administrative rules, committee transcripts) can be indexed as needed during the discovery and build phases.
Evaluation Methodology: We would maintain an evaluation set of 50-100 canonical questions with known correct answers to track accuracy and citation correctness over time. This enables regression testing when models or data sources change.
What We'll Build (Production)
Phase
Deliverable
Discovery
Conduct user interviews, validate data sources, define user personas, establish success metrics
Data Expansion
Add Oklahoma Constitution, bill summaries, fiscal impact statements
Authentication
SSO integration with state identity provider
UI Polish
Saved searches, collaboration features, export to Word
Production Hardening
Error handling, monitoring, audit logging
Training
User documentation, admin runbooks, knowledge transfer
Sample Questions It Can Answer
"What are the requirements for charter school authorizers in Title 70?"
"What does Oklahoma law say about open meetings?"
"Show me recent bills related to AI governance"
Success Metrics
Metric
Target
How Measured
Citation Precision
>95%
Citations link to correct statute/bill sections
Answer Groundedness
>90%
Answers supported by retrieved documents (no hallucination)
A multi-tenant conversational AI platform that enables any Oklahoma state agency, board, commission, or municipality to deploy an AI-powered Q&A assistant and embed it directly on their website.
Why a Unified Platform?
Standardization: Consistent chatbot quality and performance across all state entities
Statewide Observability: Centralized visibility into how citizens are engaging, what questions they're asking, and how effectively the state is answering them
Cost Control: Centralized cost center allows the state to regulate AI spending and negotiate better rates
Future-Proof: Most costs are token-based; as open-source models mature, Oklahoma can shift to locally-hosted models running on state GPUs, dramatically reducing costs at scale
Local Control: Each agency maintains full control over their documents, training data, and prompt configuration - all managed at the local level
Any Language: Multi-language support built-in; can support Spanish, Vietnamese, Korean, or any language as needed
Content Governance & Safety
Content Approval: Agency admins review and approve documents before they become searchable; draft vs. published states
PII Prevention: Automated scanning flags documents containing SSNs, phone numbers, or other PII before ingestion; admin review required
Answer Guardrails: System prompts prevent responses on topics outside agency scope; harmful content filters enabled
Escalation Path: Flagged conversations route to human review; feedback loop for continuous improvement
Accessibility: Chat widget and admin portal designed to meet WCAG 2.1 AA and Section 508 requirements
Demo for Discussion
Multi-Tenant Platform: Three demo tenants (Oklahoma City, OMES, HHS) each with isolated knowledge bases
Embeddable Chat Widget: Agencies can embed the chatbot on their existing website with their own branding
Admin Portal: Agencies upload their own documents, configure prompts, select models, and view analytics
Phased onboarding of 5-10 agencies with training and support
Key Screens
Citizen chat interface with citations
Agency admin dashboard
Document management with processing status
State-wide analytics overview
Future Capability: IVR Integration
Same AI backend can power phone IVR interactions
Seamless escalation to live agents
Call transcription and analytics
Estimated 40-60% call deflection on routine inquiries (pilot target to be validated during discovery)
Roadmap: The statewide chatbot platform is the foundation. IVR integration is the next step, leading toward a statewide contact center that serves as a centralized hub for citizen inquiries and requests - a unified help desk for all of Oklahoma state government.
Success Metrics
Metric
Target
How Measured
Deflection Rate
[40-60%]
% of inquiries resolved without human handoff (pilot target)
Citizen Satisfaction (CSAT)
>4.0/5.0
Post-conversation feedback survey
Containment Rate
[>80%]
% of conversations completed within chatbot
Time-to-Publish
[<24 hours]
Time from document upload to searchable in chatbot
An AI-powered quality assurance system that provides an additional layer of review for SNAP (food stamps) benefit cases. The system independently validates eligibility determinations and benefit calculations, helping to identify and reduce errors before they impact Oklahoma's Payment Error Rate.
Data Handling & Privacy
SNAP case data contains highly sensitive PII (SSNs, income, household composition). Our approach:
State Boundary Options: System can be deployed in AWS GovCloud or state-operated infrastructure to keep all data within state-controlled boundaries
Document Processing: Verification documents (paystubs, W-2s) are processed through Google Gemini for extraction; production deployment can use other OCR providers if required
No Training on State Data: LLM providers do not train on data sent through their APIs; we can provide contractual commitments
Encryption: All data encrypted in transit (TLS 1.3) and at rest (AES-256)
Access Controls: Role-based access; all actions logged for audit
Demo for Discussion
Case Upload Interface: Batch upload of cases (CSV) with linked verification documents
Document Extraction: AI-powered extraction of income, expenses, and household data from paystubs, W-2s, utility bills using Google Gemini
Rules Engine: Full Oklahoma SNAP FY2026 eligibility and benefit calculation rules encoded as code
Comparison Engine: Side-by-side view of Oklahoma's determination vs. CQA's independent calculation
QA Reviewer Console: Case list, detail view, findings with explanations, reviewer annotations
Analytics Dashboard: Error patterns, accuracy metrics, dollar impact, projected PER improvement
Demo Data: Synthetic case generator for demonstration purposes
V2 Preview: Demo includes early preview of V2 task management features (Processing Console, Supervisor Dashboard)
Current Architecture
What We'll Build (Production)
Phase
Deliverable
V1: QA System
Production deployment processing real (anonymized) cases alongside current QA process
Integration
Secure connection to Oklahoma's eligibility system for case data
Document Pipeline
Integration with Oklahoma's document management system
Policy Updates
Process to incorporate policy changes as rules are updated
Training Data
Capture reviewer feedback to train custom Oklahoma model
V2: Custom Model
Fine-tune AI on Oklahoma's historical QA data for higher accuracy
V3: Current™ Replacement
Full task/workflow management (if desired)
Why This Matters
Oklahoma's SNAP Payment Error Rate: 10.87% (FY2024) per USDA Food and Nutrition Service.
H.R. 1 (119th Congress, Budget Reconciliation): This federal legislation fundamentally changes the economics of SNAP administration:
Federal share of SNAP administrative costs reduces from 50% to 25%, effective FY2027 - meaning states bear 75% of administrative costs
New eligibility verification requirements with penalties for non-compliance
Stricter work requirements that increase processing complexity
National Governors Association estimates: ~$218M/year average per state in benefit cost-share implications (based on FY2024 PER data), plus an additional ~$67M/year per state for the 75% administrative cost requirement
The Opportunity: States that reduce their Payment Error Rate will avoid federal penalties and recoup costs. AI-powered QA can catch errors before benefits are issued, reducing both overpayments and underpayments while improving outcomes for Oklahoma citizens.
Future Vision: The SNAP QA system establishes the foundation for OK Benefits "Intelligent Intake" + Document Automation - extending the same AI-powered document extraction and eligibility validation to SoonerCare (Medicaid) and Child Care programs, creating a unified intelligent intake system across Oklahoma's major benefits programs.
Success Metrics
Metric
Target
How Measured
Detection Precision
>85%
% of flagged errors confirmed by QA reviewer
Detection Recall
>80%
% of actual errors caught by system (vs. ME/QC findings)
False Positive Rate
<20%
Flagged items that are not actual errors
Reviewer Time Saved
30-50%
Average case review time vs. baseline
PER Impact (projected)
1-2 point reduction
Estimated based on error detection rate and dollar impact
Project Portfolio
Alignment with Oklahoma IT Strategic Plan 2026-2028
These projects directly support the statewide strategic goals outlined by OMES:
AI-powered personalization and content recommendations for 2028 Olympics
ServiceNow Ticket Automation
5-10x productivity improvement for IT ticket handling
Legacy Forms Modernization
Modern forms platform to replace aging ok.gov sites
Occupational Licensing
Unified portal across 50+ license types
AI API Governance
Centralized gateway for statewide AI usage visibility
Digital Accessibility Compliance
AI-assisted accessibility auditing and remediation across state websites
IT Workforce Training Program
AI, cloud, and automation skill-building curriculum for state employees
Procurement AI Assistant
AI tools for drafting RFPs and solicitations for OMES and statewide procurement officials
IT Contractor Performance Assessment
AI-powered analysis of contractor deliverables, timelines, and outcomes
TravelOK Analytics Modernization
Server-side tracking, unified attribution dashboard, and UX analytics for Oklahoma Tourism
Engagement Framework
How We Approach Every Project
Each engagement follows a structured process designed to minimize risk, maximize learning, and deliver working software quickly.
┌─────────────────────────────────────────────────────────────────────┐
│ 1. SCOPING MEETINGS (1 week) │
│ • Understand the problem and stakeholders │
│ • Review existing systems and data │
│ • Identify quick wins vs. complex dependencies │
│ • Align on success criteria │
├─────────────────────────────────────────────────────────────────────┤
│ 2. DISCOVERY PERIOD (1-2 weeks) │
│ • Deep-dive into current processes and pain points │
│ • Data inventory and integration mapping │
│ • Technical architecture assessment │
│ • Security and compliance requirements │
│ • Detailed scope and timeline development │
├─────────────────────────────────────────────────────────────────────┤
│ 2.5 SECURITY REVIEW (parallel with build) │
│ • Architecture review with state security team │
│ • Security artifacts delivered (SSP inputs, data flows) │
│ • ATO preparation begins during discovery │
│ • Penetration testing support if required │
│ • Security approval gate before production rollout │
├─────────────────────────────────────────────────────────────────────┤
│ 3. BUILD & ITERATE (varies by project) │
│ • Weekly (or more frequent) stakeholder meetings │
│ • Live demos of working software each session │
│ • Rapid iteration based on feedback │
│ • Continuous integration with state systems │
│ • OMES staff embedded for knowledge transfer │
├─────────────────────────────────────────────────────────────────────┤
│ 4. SOFT LAUNCH / BETA (2-4 weeks) │
│ • Production deployment with limited users │
│ • Real usage without cutting over existing systems │
│ • Bug fixes and refinements based on live feedback │
│ • Performance monitoring and optimization │
│ • Training for broader rollout │
├─────────────────────────────────────────────────────────────────────┤
│ 5. LAUNCH / ROLLOUT (staged) │
│ • Phased rollout by agency, region, or user group │
│ • Parallel operation with legacy systems if needed │
│ • Full monitoring and support │
│ • Success metrics tracking │
│ • Handoff to ongoing operations │
└─────────────────────────────────────────────────────────────────────┘
Pricing Model
Annual Platform License (SaaS Model)
All projects will be priced as annual platform licenses rather than time-and-materials or fixed-price project fees. This approach:
Aligns incentives: We succeed when the software delivers value, not when we bill hours
Includes everything: Implementation, iteration, support, and enhancements
Predictable budgeting: Fixed annual cost for planning purposes
Continuous improvement: Platform evolves based on user feedback
No surprise costs: Scope changes handled within the license
AI model costs: LLM token usage is a passthrough cost billed at actual usage rates
Detailed pricing to be developed based on specific scope and requirements.
Engineering Pod Composition
Each project is delivered by a dedicated "engineering pod" - a small, focused team that works exclusively on your initiative:
1
Product Lead / Architect
2-3
Product Engineers
1
AI/ML Engineer
1
Product Manager
1
OMES Liaison
The OMES liaison is an embedded state employee who works alongside the pod, ensuring knowledge transfer and building internal capability throughout the engagement.
Cadence: Weekly demos, backlog grooming, stakeholder reviews, and sprint planning ensure continuous alignment and visibility.
Artifacts Delivered: Code repository, Infrastructure as Code (IaC), operational runbooks, admin guides, training sessions, and knowledge transfer plan.
Security, Privacy & Compliance
Security and compliance are foundational to every engagement. We build systems that meet state security requirements and work alongside your security teams throughout the process.
Data Classification & Handling
Classification
Examples
Handling
Public
Published statutes, public website content, press releases
Standard cloud hosting; encrypted in transit and at rest
Strict access controls; field-level encryption; comprehensive audit trails; data minimization
Deployment Options
GovCloud: AWS GovCloud or Azure Government for workloads requiring enhanced compliance controls
Commercial Cloud with Controls: Standard cloud infrastructure with SOC 2 compliance, data residency agreements, and state-approved security configurations
Our architecture is designed to be portable across cloud providers. Technology choices (databases, vector stores, hosting) are implementation details that can be adjusted based on state requirements and existing contracts.
Audit Logging & Retention
What's Logged: User authentication, data access, queries, admin actions, system events, and configuration changes
Retention: Configurable to match state policy (typically 1-7 years)
Access: Logs accessible to authorized state personnel; exportable for compliance reviews
Storage: Logs stored separately from application data with tamper-evident controls
AI Model Data Handling
What goes to LLM providers: User queries and relevant document chunks for context generation
What never leaves: Raw PII, complete case files, credentials, or bulk data exports
Data protection: No data is used to train external models and all data is held within the USA
ATO Readiness
We provide security artifacts to support your Authority to Operate (ATO) process:
System Security Plan (SSP) inputs
Architecture and data flow diagrams
Threat model documentation
Penetration test support and remediation
Incident response procedures
Vulnerability management approach
We work alongside state security teams throughout the engagement - security review is a collaborative process, not a gate at the end.
Next Steps
January 23, 4:30pm: Whiteboard session at Lincoln Data Center
Review working demos together
Align on priority projects (these three, others, or different sequencing)
Identify key stakeholders for each initiative
Discuss pricing framework
Post-Meeting:
Schedule discovery calls for priority projects
Gather stakeholder requirements
Finalize per-project pricing
Appendix A: Legislative Research Assistant Discovery Questions
Data & Sources
Is LegiScan data sufficient for Oklahoma statutes, or do you need authoritative oklegislature.gov text?
LegiScan provides structured data but may lag by days
oklegislature.gov is authoritative but requires scraping
Are bill summaries and fiscal impact statements consistently available for all bills?
These are high-value for research; need to confirm coverage
Do you need "Session in Review" documents or other staff publications indexed?
Referenced on okhouse.gov as valuable research resources
Are there internal documents (memos, talking points, templates) that should be searchable?
Would require secure handling; affects architecture
Do you need case law, AG opinions, or administrative rules in scope?