Skip to main content

AI Cross-Validation

AI Cross-Validation is a quality assurance feature in Wegent that uses another AI model to evaluate and correct agent responses, ensuring accuracy, logic, and completeness.


πŸ“‹ Table of Contents​


βœ… What is AI Cross-Validation​

AI Cross-Validation is a dual-verification mechanism. When enabled, the system uses another AI model you select to evaluate the agent's response, checking accuracy, logic, and completeness, and providing improvement suggestions when issues are found.

Workflow:

Agent response β†’ Validation model evaluates β†’ Generates scores and suggestions β†’ Provides improved version β†’ User chooses whether to apply

Core Benefits​

BenefitDescription
Quality AssuranceVerify response quality through a second model
Error DetectionIdentify factual errors and logical gaps
Improvement SuggestionsProvide specific correction suggestions
Optional ApplicationUsers can choose whether to adopt corrections

πŸš€ Enabling AI Cross-Validation​

Step 1: Click the Cross-Validation Button​

  1. Find the cross-validation icon (βœ“ checkmark circle) in the chat input area
  2. Click the icon to open the model selection dialog

Step 2: Select Validation Model​

In the popup dialog:

  1. Search Models: Use the search box to filter models
  2. Browse List: View available models
  3. Select Model: Click to select the model for cross-validation
  4. Confirm Selection: AI cross-validation is automatically enabled after selection

Step 3: Confirm Enabled Status​

  • Cross-validation icon highlights when enabled
  • Hover to see the current validation model name
  • Click again to disable AI cross-validation

Model Types​

TypeDescription
Public ModelsSystem pre-configured models
User ModelsUser-defined custom models

State Persistence​

  • AI cross-validation settings are saved to local storage
  • Settings are automatically restored when switching tasks
  • New tasks inherit previous settings

πŸ“Š Understanding Evaluation Results​

After the agent responds, the validation model generates evaluation results displayed in a panel below the response.

Scoring Metrics​

The validation model evaluates responses across three dimensions:

MetricDescriptionScore Range
AccuracyAre facts correct0-10
LogicIs reasoning sound0-10
CompletenessIs the answer comprehensive0-10

Score Color Indicators​

Score RangeColorMeaning
8-10GreenExcellent
6-7YellowGood
4-5OrangeNeeds Improvement
0-3RedSignificant Issues

Issue List​

If issues are found, specific problems and suggestions are displayed:

  • Issue Description: Points out specific problems in the response
  • Correction Suggestion: Provides suggestions for improvement
  • Issue Number: Numbered sequentially for reference

Improved Version​

The validation model generates an improved version of the response:

  • Displayed in the main area of the evaluation panel
  • Issues found have been corrected
  • Maintains the style and structure of the original response

Summary​

A brief summary at the bottom of evaluation results provides an overview of the overall assessment.


βœ… Applying Corrections​

View Improved Version​

  1. After evaluation completes, the improved version displays at the top of the panel
  2. Compare the original response with the improved version
  3. Click "Show Evaluation Details" to see specific scores and issues

Apply Improvements​

If you approve the improved version:

  1. Hover over the improved version area
  2. Click the "Apply" button in the top right
  3. The improved version replaces the original response
  4. Button changes to "Applied" status

Application Status​

StatusButton DisplayDescription
Not Applied"Apply"Can click to apply improvement
ApplyingLoading animationSaving improvement
Applied"Applied" βœ“Improvement has been adopted

Re-evaluate​

If re-evaluation is needed:

  1. Expand evaluation details
  2. Click the "Re-validate" button at the bottom
  3. Validation model will re-analyze the response

🎯 Use Cases​

Case 1: Fact Checking​

When responses involve factual information:

  • Validation model verifies accuracy of facts
  • Identifies possible errors or outdated information
  • Provides more accurate information version

Case 2: Logic Verification​

When responses involve reasoning or analysis:

  • Checks if reasoning process is sound
  • Identifies logical gaps or contradictions
  • Provides more rigorous argumentation

Case 3: Completeness Check​

When responses need comprehensive coverage:

  • Checks if important aspects are missing
  • Identifies content that needs supplementation
  • Provides more complete answers

Case 4: Important Decisions​

When responses are used for important decisions:

  • Provides a second opinion
  • Increases response credibility
  • Reduces risk of wrong decisions

✨ Best Practices​

1. Choosing the Right Validation Model​

Recommended Strategies:

  • Choose a different model from the main model for different perspectives
  • For specialized domains, choose models that perform well in that area
  • Consider model speed and cost

2. When to Enable AI Cross-Validation​

Recommended to Enable:

  • Responses involving factual information
  • Important decision support
  • Scenarios requiring high accuracy
  • Professional domain Q&A

Can Skip:

  • Simple creative writing
  • Non-critical conversations
  • Scenarios requiring fast responses

3. How to Use Evaluation Results​

  • Check Scores: Quickly understand response quality
  • Read Issues: Understand specific improvement points
  • Compare Versions: Compare original and improved versions
  • Selective Application: Decide whether to apply based on actual needs

4. Combining with Other Features​

  • With Smart Follow-up Mode: First clarify requirements, then verify with cross-validation
  • With Knowledge Bases: Responses based on knowledge bases are easier to verify
  • Iterative Optimization: Optimize subsequent questions based on validation feedback

⚠️ Common Issues​

Q1: Does AI cross-validation increase response time?​

Answer: Yes, cross-validation requires additional model calls, adding some wait time. However, for scenarios requiring high-quality responses, this wait is worthwhile.

Q2: Is the validation model's evaluation always accurate?​

Answer: The validation model's evaluation is a reference opinion, not necessarily 100% accurate. It's recommended to combine with your own judgment when deciding whether to adopt.

Q3: Can I use the same model for cross-validation?​

Answer: Technically possible, but using a different model is recommended for different perspectives and more effective verification.

Q4: Can I undo after applying correction?​

Answer: After application, the original response is replaced and cannot be directly undone. It's recommended to carefully compare both versions before applying.

Q5: Which agents support AI cross-validation?​

Answer: AI cross-validation currently only supports Chat Shell type agents.

Q6: Are cross-validation settings saved?​

Answer: Yes, AI cross-validation settings (including selected model) are saved to local storage and automatically restored when accessing the same task next time.

Q7: Why is there sometimes no improved version?​

Answer: If the validation model determines the original response is good enough, it may not generate an improved version, only showing scores and a "no correction needed" message.



Use AI Cross-Validation to ensure AI response quality! βœ