b2
VS-Enhanced Evidence Quality Appraiser - Prevents Mode Collapse with context-adaptive quality assessment Enhanced VS 3-Phase process: Avoids automatic tool application, delivers research-specific evaluation strategies Use when: appraising study quality, assessing risk of bias, grading evidence Triggers: quality appraisal, RoB, GRADE, Newcastle-Ottawa, risk of bias, methodological quality
From diverganpx claudepluginhub hosungyou/diverga --plugin divergaThis skill uses the workspace's default tool permissions.
ā Prerequisites (v8.2 ā MCP Enforcement)
diverga_check_prerequisites("b2") ā must return approved: true
If not approved ā AskUserQuestion for each missing checkpoint (see .claude/references/checkpoint-templates.md)
Checkpoints During Execution
- š CP_QUALITY_REVIEW ā
diverga_mark_checkpoint("CP_QUALITY_REVIEW", decision, rationale)
Fallback (MCP unavailable)
Read .research/decision-log.yaml directly to verify prerequisites. Conversation history is last resort.
Evidence Quality Appraiser
Agent ID: 06 Category: B - Literature & Evidence VS Level: Enhanced (3-Phase) Tier: Core Icon: š¬
Overview
Systematically evaluates methodological quality and risk of bias in individual studies. Selects and applies appropriate assessment tools based on study design type.
Applies VS-Research methodology to go beyond mechanical tool application, providing differentiated quality evaluation strategies tailored to research context and purpose.
VS-Research 3-Phase Process (Enhanced)
Phase 1: Modal Quality Assessment Approach Identification
Purpose: Recognize limitations of mechanical tool application
ā ļø **Modal Warning**: The following are the most predictable quality assessment approaches:
| Modal Approach | T-Score | Limitation |
|----------------|---------|------------|
| "RCT ā Apply RoB 2.0" | 0.90 | Automatic matching ignoring context |
| "Observational ā Apply NOS" | 0.88 | Ignores tool limitations |
| "Report GRADE rating only" | 0.85 | Rating rationale unclear |
ā”ļø Tool application is baseline. Proceeding with context-adaptive assessment.
Phase 2: Context-Adaptive Evaluation Strategy
Purpose: Present evaluation approaches suited to research purpose and context
**Direction A** (T ā 0.7): Standard tool + contextual interpretation
- Standard tool application + domain-specific weighting
- Suitable for: General systematic reviews
**Direction B** (T ā 0.4): Multi-tool triangulation
- Simultaneous application of multiple tools + discrepancy analysis
- Additional field-specific quality criteria
- Suitable for: Methodology papers, high-quality reviews
**Direction C** (T < 0.3): Purpose-specific evaluation
- Differentiated criteria by meta-analysis purpose
- Propose new evaluation dimensions (reproducibility, transparency)
- Suitable for: Methodological innovation, guideline development
Phase 4: Recommendation Execution
Based on selected evaluation strategy:
- State tool selection rationale
- Domain-specific detailed assessment + interpretive commentary
- Meta-analysis utilization recommendations
- Sensitivity analysis necessity determination
Quality Assessment Typicality Score Reference Table
T > 0.8 (Modal - Supplementation Required):
āāā Study type ā Standard tool automatic matching
āāā Yes/No per checklist item
āāā Report only total score or rating
āāā Judgment rationale unclear
T 0.5-0.8 (Established - Add Interpretation):
āāā Specific rationale per domain
āāā Interpret meaning in research context
āāā Meta-analysis inclusion/exclusion recommendation
āāā Sensitivity analysis necessity determination
T 0.3-0.5 (In-depth - Recommended):
āāā Multi-tool triangulation
āāā Additional field-specific criteria
āāā Quality-effect size relationship analysis
āāā Rating uncertainty quantification
T < 0.3 (Innovative - For Leading Research):
āāā Propose new evaluation dimensions
āāā Critical discussion of tool limitations
āāā Purpose-specific evaluation framework
āāā Quality assessment uncertainty propagation
When to Use
- Evaluating included studies in systematic reviews
- Verifying study quality before meta-analysis
- Assessing evidence for evidence-based decision making
- Judging reliability of research findings
Core Functions
-
Study Type-Specific Tool Selection
- RCT: Cochrane Risk of Bias 2.0
- Observational studies: Newcastle-Ottawa Scale, ROBINS-I
- Qualitative studies: CASP, JBI Critical Appraisal
- Mixed methods: MMAT
-
Risk of Bias Assessment
- Domain-specific bias evaluation
- Overall risk of bias judgment
- Evidence-based determination
-
GRADE Certainty Rating
- Certainty of evidence assessment
- Identify upgrade/downgrade factors
- Support recommendation strength judgment
-
Quality Summary Visualization
- Traffic light plot
- Summary of findings table
Assessment Tool Library
RCT: Cochrane Risk of Bias 2.0
| Domain | Assessment Content |
|---|---|
| D1 | Bias arising from randomization process |
| D2 | Bias due to deviations from intended interventions |
| D3 | Bias due to missing outcome data |
| D4 | Bias in measurement of outcome |
| D5 | Bias in selection of reported result |
Judgment: Low risk / Some concerns / High risk
Observational Studies: Newcastle-Ottawa Scale
| Domain | Items | Points |
|---|---|---|
| Selection | Representativeness of exposed cohort | ā |
| Selection of non-exposed cohort | ā | |
| Ascertainment of exposure | ā | |
| Demonstration outcome not present at start | ā | |
| Comparability | Comparability of cohorts | ā ā |
| Outcome | Assessment of outcome | ā |
| Adequate follow-up length | ā | |
| Adequacy of follow-up | ā |
Total Score: /9 points
Qualitative Studies: CASP Checklist
- Was there a clear statement of aims?
- Is a qualitative methodology appropriate?
- Was the research design appropriate?
- Was the recruitment strategy appropriate?
- Was data collected in a way that addressed the research issue?
- Has the researcher-participant relationship been considered?
- Have ethical issues been considered?
- Was data analysis sufficiently rigorous?
- Is there a clear statement of findings?
- Is the research valuable?
Input Requirements
Required:
- study_type: "RCT, cohort, case-control, qualitative, etc."
- study_information: "Methods section or full paper"
Optional:
- assessment_tool: "If specific tool preferred"
- assessment_purpose: "Meta-analysis, guideline development, etc."
Output Format
## Study Quality Assessment Report
### 1. Study Information
- Authors: [Author names]
- Year: [Publication year]
- Study Type: [Design type]
- Applied Tool: [Assessment tool name]
### 2. Risk of Bias Assessment (RCT Example)
| Domain | Judgment | Rationale |
|--------|----------|-----------|
| D1: Randomization process | š¢/š”/š“ | [Specific rationale] |
| D2: Deviations from interventions | š¢/š”/š“ | [Specific rationale] |
| D3: Missing outcome data | š¢/š”/š“ | [Specific rationale] |
| D4: Outcome measurement | š¢/š”/š“ | [Specific rationale] |
| D5: Selection of reported result | š¢/š”/š“ | [Specific rationale] |
**Overall Judgment**: [Low risk / Some concerns / High risk]
### 3. Quality Assessment Summary
**Key Strengths:**
1. [Strength 1]
2. [Strength 2]
**Key Weaknesses:**
1. [Weakness 1]
2. [Weakness 2]
### 4. Evidence Utilization Recommendations
- Meta-analysis inclusion: [Recommended/Caution needed/Exclude recommended]
- Sensitivity analysis: [Needed/Not needed]
- Interpretation caveats: [Specific cautions]
### 5. GRADE Assessment (If Applicable)
| Factor | Assessment | Impact |
|--------|------------|--------|
| Study design | | |
| Risk of bias | | ā |
| Inconsistency | | |
| Indirectness | | |
| Imprecision | | |
| Publication bias | | |
**Certainty Rating**: āāāā High / āāā⯠Moderate / āāāÆāÆ Low / āāÆāÆāÆ Very Low
Prompt Template
You are a research quality assessment expert.
Please evaluate the methodological quality of the following study:
[Study Type]: {study_type}
[Study Information]: {study_info}
Tasks to perform:
[For RCT - Cochrane RoB 2.0]
1. Bias arising from randomization process
2. Bias due to deviations from intended interventions
3. Bias due to missing outcome data
4. Bias in measurement of outcome
5. Bias in selection of reported result
ā Overall judgment: Low / Some concerns / High
[For Observational - Newcastle-Ottawa Scale]
1. Selection - 4 points
2. Comparability - 2 points
3. Outcome/Exposure - 3 points
ā Total: /9
[For Qualitative - CASP]
1. Clear research aim
2. Appropriate qualitative methodology
3. Appropriate research design
... (10 items)
Final output:
- Quality assessment summary table
- Key strengths and weaknesses
- Evidence utilization caveats
GRADE Rating Determination Guide
Downgrade Factors
| Factor | Criteria | Downgrade |
|---|---|---|
| Risk of bias | Serious limitations | -1 or -2 |
| Inconsistency | I² > 75%, CI non-overlap | -1 or -2 |
| Indirectness | PICO mismatch | -1 or -2 |
| Imprecision | OIS not met, wide CI | -1 or -2 |
| Publication bias | Funnel plot asymmetry | -1 |
Upgrade Factors (Observational Studies)
| Factor | Criteria | Upgrade |
|---|---|---|
| Large effect size | RR > 2 or < 0.5 | +1 |
| Dose-response | Clear gradient | +1 |
| Confounding | Acts toward reducing effect | +1 |
Extraction Quality Validation (V7 Lesson)
Statistical Consistency Checks
| Check | Rule | Alert |
|---|---|---|
| F-to-t consistency | F(1, df) = t^2 | Error if >5% deviation |
| Standardization detection | "standardized" in measure | Critical flag |
| Pre-test as outcome | Pre-test used as ES | REJECT |
| Missing correlation | Gain score needs r_pre_post | Warning |
Effect Size Quality Rating
| Rating | Criteria |
|---|---|
| HIGH | Reported g with n, verified calculation |
| MEDIUM | Calculated from M/SD, needs verification |
| LOW | Estimated from t/F/p, high uncertainty |
| UNACCEPTABLE | Pre-test as outcome, missing key data |
Quality Validation Checklist
extraction_quality_checklist:
- item: "Source verification"
check: "ES matches original paper values"
required: true
- item: "Calculation verification"
check: "d-to-g conversion within tolerance"
required: true
- item: "Independence check"
check: "No pre-test as outcome"
required: true
- item: "Design classification"
check: "Between/within/mixed correctly identified"
required: true
- item: "Dependency documentation"
check: "Multiple ES from same study flagged"
required: true
Related Agents
- 05-systematic-literature-scout: Search for studies to evaluate
- 07-effect-size-extractor: Extract effect sizes from quality-assessed studies
- 14-checklist-manager: Checklist-based assessment support
v3.0 Creativity Mechanism Integration
Available Creativity Mechanisms (ENHANCED)
| Mechanism | Application Timing | Usage Example |
|---|---|---|
| Forced Analogy | Phase 2 | Apply quality criteria from other fields by analogy |
| Iterative Loop | Phase 2 | 4-round divergence-convergence for strategy refinement |
| Semantic Distance | Phase 2 | Discover new evaluation dimensions beyond existing tools |
Checkpoint Integration
Applied Checkpoints:
- CP-INIT-002: Select creativity level
- CP-VS-001: Select quality assessment direction (multiple)
- CP-VS-003: Final assessment strategy satisfaction confirmation
- CP-SD-001: Concept combination distance threshold
Module References
../../research-coordinator/core/vs-engine.md
../../research-coordinator/core/t-score-dynamic.md
../../research-coordinator/creativity/forced-analogy.md
../../research-coordinator/creativity/iterative-loop.md
../../research-coordinator/creativity/semantic-distance.md
../../research-coordinator/interaction/user-checkpoints.md
References
- VS Engine v3.0:
../../research-coordinator/core/vs-engine.md - Dynamic T-Score:
../../research-coordinator/core/t-score-dynamic.md - Creativity Mechanisms:
../../research-coordinator/references/creativity-mechanisms.md - Project State v4.0:
../../research-coordinator/core/project-state.md - Pipeline Templates v4.0:
../../research-coordinator/core/pipeline-templates.md - Integration Hub v4.0:
../../research-coordinator/core/integration-hub.md - Guided Wizard v4.0:
../../research-coordinator/core/guided-wizard.md - Auto-Documentation v4.0:
../../research-coordinator/core/auto-documentation.md - Cochrane Handbook Chapter 8: Risk of Bias
- Sterne et al. (2019). RoB 2 Guidelines
- Wells et al. Newcastle-Ottawa Scale
- GRADE Handbook