Multi-Provider Verification
TruthVouch simultaneously monitors ChatGPT, Claude, Gemini, Perplexity, and 5+ other AI engines. Multi-provider verification normalizes responses across different model styles and outputs.
Supported Providers
| Provider | Models | Status |
|---|---|---|
| OpenAI | GPT-4, GPT-4 Turbo, GPT-3.5 Turbo | Active |
| Anthropic | Claude 3 Opus, Sonnet, Haiku | Active |
| Gemini Pro, Ultra, Flash | Active | |
| Perplexity | Custom model | Active |
| Cohere | Command | Active |
| Mistral | Large, Medium | Active |
| Meta | Llama 2, 3 | Active |
| Azure OpenAI | Azure-hosted OpenAI | Active |
| AWS Bedrock | Multi-model | Active |
Query Normalization
Different LLMs respond with different styles and phrasing to the same question. TruthVouch normalizes these responses to extract consistent, comparable claims:
- Same query across all providers produces comparable outputs
- Entity and claim extraction identifies key factual assertions
- Responses are normalized to standard format for verification
- Confidence scores reflect the modelβs stated certainty
This normalization enables fair, apples-to-apples comparison across all 9+ supported AI engines.
How It Works
TruthVouch sends the same canonical query to all 9+ supported providers simultaneously. Each response is normalized into a standard format, then verified against your truth nuggets. This parallel multi-provider approach catches hallucinations that might slip through a single AI engine.
Response Aggregation
When you run a cross-check query, TruthVouch aggregates responses to show:
- What each provider said
- Consensus answer (when providers agree)
- Agreement rate (percentage in alignment)
- Overall confidence in the result
Conflict Detection
Detects when providers disagree:
ChatGPT: "Founded in 2023"Claude: "Founded in 2023"Gemini: "Founded in 2024"Perplexity: "Founded in 2023"
Conflict Detected: Agreement: 3/4 (75%) Outlier: Gemini (likely hallucinating) Recommendation: Alert, require manual reviewProvider Reliability Scoring
Track which providers are most reliable for your specific content:
- Accuracy: Percentage of correct responses
- Hallucination rate: Percentage with detected hallucinations
- Response consistency: How often responses agree with other providers
Dashboard View
Monitor all 9+ providers in one view:
- Navigate to Shield β Multi-Provider Monitoring
- See:
- Heat map of which claims each provider gets right/wrong
- Provider reliability trends
- Cross-provider consensus scores
- Agreement analysis
Configuration
You can enable or disable specific providers and set consensus thresholds in Dashboard β Monitoring Settings:
- Select which providers to monitor
- Set check frequency
- Configure alerts when providers disagree
- Set minimum consensus threshold for alerts
Next Steps
- Multi-Provider Monitoring: Set up provider tracking
- Conflict Alerts: Get notified when providers disagree
- Provider Analytics: Compare provider reliability over time
- Dashboard: Monitor all 9+ providers in one view