The platform is provider-agnostic. Enable whichever LLM adapters you want by setting the matching environment variables, or use the runtime API key box inside a project workspace when you want to test AI actions without rebuilding the app.
This is the default runtime path for ingestion, page previews, and baseline document parsing.
pdf-ocr-cvConfiguredLocal runtime
Best at
- PDF splitting and page indexing
- Text extraction and table capture
- Deterministic fallback when no LLM provider is configured
Default models
PyMuPDFpdfplumberPillow
Recommended for document-grounded reasoning and estimator-facing explanations once configured.
multimodal-llmNot configuredAPI key
Best at
- Multimodal drawing and specification reasoning
- Evidence-grounded ambiguity explanation
- Agent orchestration for scope extraction
Default models
gpt-5gpt-4.1
Environment variable
OPENESTIMATOR_OPENAI_API_KEY
Positioned as an interchangeable multimodal provider for page reasoning and benchmark workflows.
multimodal-llmNot configuredAPI key
Best at
- Large drawing-set summarization
- Fast sheet-level triage
- Multimodal page and schedule understanding
Default models
gemini-2.5-progemini-2.5-flash
Environment variable
OPENESTIMATOR_GEMINI_API_KEY
Useful for disciplined trade-scope reasoning once credentials are provided.
llmNot configuredAPI key
Best at
- Long-context specification reasoning
- Assumption and exclusion drafting
- Estimator-facing explanation quality
Default models
claude-sonnetclaude-opus
Environment variable
OPENESTIMATOR_ANTHROPIC_API_KEY
This is the modular slot for computer-vision-assisted count, measurement, and markup extraction.
cv-geometryConfiguredLocal runtime
Best at
- Plan element extraction scaffolding
- Future count and geometry assist modules
- Legend and symbol alignment candidates
Default models
linework calibrationsymbol spottingregion proposal