AI Cross-Validation
AI Cross-Validation is a quality assurance feature in Wegent that uses another AI model to evaluate and correct agent responses, ensuring accuracy, logic, and completeness.
π Table of Contentsβ
- What is AI Cross-Validation
- Enabling AI Cross-Validation
- Understanding Evaluation Results
- Applying Corrections
- Use Cases
- Best Practices
- Common Issues
β What is AI Cross-Validationβ
AI Cross-Validation is a dual-verification mechanism. When enabled, the system uses another AI model you select to evaluate the agent's response, checking accuracy, logic, and completeness, and providing improvement suggestions when issues are found.
Workflow:
Agent response β Validation model evaluates β Generates scores and suggestions β Provides improved version β User chooses whether to apply
Core Benefitsβ
| Benefit | Description |
|---|---|
| Quality Assurance | Verify response quality through a second model |
| Error Detection | Identify factual errors and logical gaps |
| Improvement Suggestions | Provide specific correction suggestions |
| Optional Application | Users can choose whether to adopt corrections |
π Enabling AI Cross-Validationβ
Step 1: Click the Cross-Validation Buttonβ
- Find the cross-validation icon (β checkmark circle) in the chat input area
- Click the icon to open the model selection dialog
Step 2: Select Validation Modelβ
In the popup dialog:
- Search Models: Use the search box to filter models
- Browse List: View available models
- Select Model: Click to select the model for cross-validation
- Confirm Selection: AI cross-validation is automatically enabled after selection
Step 3: Confirm Enabled Statusβ
- Cross-validation icon highlights when enabled
- Hover to see the current validation model name
- Click again to disable AI cross-validation
Model Typesβ
| Type | Description |
|---|---|
| Public Models | System pre-configured models |
| User Models | User-defined custom models |
State Persistenceβ
- AI cross-validation settings are saved to local storage
- Settings are automatically restored when switching tasks
- New tasks inherit previous settings
π Understanding Evaluation Resultsβ
After the agent responds, the validation model generates evaluation results displayed in a panel below the response.
Scoring Metricsβ
The validation model evaluates responses across three dimensions:
| Metric | Description | Score Range |
|---|---|---|
| Accuracy | Are facts correct | 0-10 |
| Logic | Is reasoning sound | 0-10 |
| Completeness | Is the answer comprehensive | 0-10 |
Score Color Indicatorsβ
| Score Range | Color | Meaning |
|---|---|---|
| 8-10 | Green | Excellent |
| 6-7 | Yellow | Good |
| 4-5 | Orange | Needs Improvement |
| 0-3 | Red | Significant Issues |
Issue Listβ
If issues are found, specific problems and suggestions are displayed:
- Issue Description: Points out specific problems in the response
- Correction Suggestion: Provides suggestions for improvement
- Issue Number: Numbered sequentially for reference
Improved Versionβ
The validation model generates an improved version of the response:
- Displayed in the main area of the evaluation panel
- Issues found have been corrected
- Maintains the style and structure of the original response
Summaryβ
A brief summary at the bottom of evaluation results provides an overview of the overall assessment.
β Applying Correctionsβ
View Improved Versionβ
- After evaluation completes, the improved version displays at the top of the panel
- Compare the original response with the improved version
- Click "Show Evaluation Details" to see specific scores and issues
Apply Improvementsβ
If you approve the improved version:
- Hover over the improved version area
- Click the "Apply" button in the top right
- The improved version replaces the original response
- Button changes to "Applied" status
Application Statusβ
| Status | Button Display | Description |
|---|---|---|
| Not Applied | "Apply" | Can click to apply improvement |
| Applying | Loading animation | Saving improvement |
| Applied | "Applied" β | Improvement has been adopted |
Re-evaluateβ
If re-evaluation is needed:
- Expand evaluation details
- Click the "Re-validate" button at the bottom
- Validation model will re-analyze the response
π― Use Casesβ
Case 1: Fact Checkingβ
When responses involve factual information:
- Validation model verifies accuracy of facts
- Identifies possible errors or outdated information
- Provides more accurate information version
Case 2: Logic Verificationβ
When responses involve reasoning or analysis:
- Checks if reasoning process is sound
- Identifies logical gaps or contradictions
- Provides more rigorous argumentation
Case 3: Completeness Checkβ
When responses need comprehensive coverage:
- Checks if important aspects are missing
- Identifies content that needs supplementation
- Provides more complete answers
Case 4: Important Decisionsβ
When responses are used for important decisions:
- Provides a second opinion
- Increases response credibility
- Reduces risk of wrong decisions
β¨ Best Practicesβ
1. Choosing the Right Validation Modelβ
Recommended Strategies:
- Choose a different model from the main model for different perspectives
- For specialized domains, choose models that perform well in that area
- Consider model speed and cost
2. When to Enable AI Cross-Validationβ
Recommended to Enable:
- Responses involving factual information
- Important decision support
- Scenarios requiring high accuracy
- Professional domain Q&A
Can Skip:
- Simple creative writing
- Non-critical conversations
- Scenarios requiring fast responses
3. How to Use Evaluation Resultsβ
- Check Scores: Quickly understand response quality
- Read Issues: Understand specific improvement points
- Compare Versions: Compare original and improved versions
- Selective Application: Decide whether to apply based on actual needs
4. Combining with Other Featuresβ
- With Smart Follow-up Mode: First clarify requirements, then verify with cross-validation
- With Knowledge Bases: Responses based on knowledge bases are easier to verify
- Iterative Optimization: Optimize subsequent questions based on validation feedback
β οΈ Common Issuesβ
Q1: Does AI cross-validation increase response time?β
Answer: Yes, cross-validation requires additional model calls, adding some wait time. However, for scenarios requiring high-quality responses, this wait is worthwhile.
Q2: Is the validation model's evaluation always accurate?β
Answer: The validation model's evaluation is a reference opinion, not necessarily 100% accurate. It's recommended to combine with your own judgment when deciding whether to adopt.
Q3: Can I use the same model for cross-validation?β
Answer: Technically possible, but using a different model is recommended for different perspectives and more effective verification.
Q4: Can I undo after applying correction?β
Answer: After application, the original response is replaced and cannot be directly undone. It's recommended to carefully compare both versions before applying.
Q5: Which agents support AI cross-validation?β
Answer: AI cross-validation currently only supports Chat Shell type agents.
Q6: Are cross-validation settings saved?β
Answer: Yes, AI cross-validation settings (including selected model) are saved to local storage and automatically restored when accessing the same task next time.
Q7: Why is there sometimes no improved version?β
Answer: If the validation model determines the original response is good enough, it may not generate an improved version, only showing scores and a "no correction needed" message.
π Related Resourcesβ
- Creating Conversations - Learn how to create conversations
- Smart Follow-up Mode - Learn about smart follow-up mode
- Configuring Models - Add and configure models
Use AI Cross-Validation to ensure AI response quality! β