method-gap-detector
Detects methodological gaps across study design, analysis, validation, bias control, reproducibility, and implementation readiness within a biomedical research area. Use this skill when a user wants to identify what current studies are still methodologically missing, which weaknesses are most consequential, and what upgrade path would produce a stronger next-step study. Always separate design gaps, analysis gaps, validation gaps, and reproducibility gaps. Never treat technical complexity as methodological rigor.
Veto GatesRequired pass for any deployment consideration
| Dimension | Result | Detail |
|---|---|---|
| Scientific Integrity | PASS | Hard Rule #11 explicitly prohibits fabricating references, PMIDs, DOIs, software details, validation claims, cohort properties, or study findings. No fabricated data detected. |
| Practice Boundaries | PASS | Explicit out-of-scope redirect for patient-specific treatment advice and live statistical consulting. No clinical recommendations issued. |
| Methodological Ground | PASS | Hard Rules #3-4 (technical complexity ≠ rigor; internal validation ≠ external validation) prevent the most common methodological fallacies. Hard Rule #13 mandates uncertainty labeling. Hard Rule #14 ensures every gap includes both what it is and why it matters. |
| Code Usability | N/A | Mode A direct execution — no code generated. |
Core Capability90 / 100 — 8 Categories
Medical TaskExecution Average: 84.4 / 100 — Assertions: 32/35 Passed
Full A-I output produced. Four-category gap classification applied. Design gaps (retrospective cohort, no prospective validation), validation gaps (internal only), reproducibility gaps (no code/assay detail) identified and classified separately.
Batch effect gap (analysis), cell clustering variability (reproducibility), and absent clinical outcome validation (validation) identified and classified. Technical sophistication not conflated with rigor.
Feature instability (analysis), inter-scanner variability (transportability), lack of prospective validation (validation), and harmonization absence (reproducibility) identified and classified separately.
Step 1 narrowing applied. Assessment proceeds for a narrowed topic unit. Cannot assess all cancer biomarker research as a single unit.
Cross-paper pattern detection applied. Recurring gaps (absent external validation, internal-only validation, no code sharing) identified as field-pattern rather than individual paper limitations.
Out-of-scope redirect correctly issued. No statistical analysis of live data attempted. Adjacent in-scope alternative offered.
Hard Rule #3 correctly applied. Technical sophistication not equated with methodological rigor. Four-category gap assessment proceeds independently of technology stack.
Key Strengths
- Four-category methodological gap taxonomy (design/analysis/validation/reproducibility) provides systematic, non-collapsing coverage — directly prevents the most common failure of vague limitations lists
- Hard Rule #3 ('technical complexity ≠ methodological rigor') is a rare and important safeguard against the widespread conflation of sophisticated methods with valid methodology
- Hard Rules #7-8 (prioritize consequential over common gaps; upgrade must target identified weakness) enforce tight reasoning chains from gap identification to upgrade recommendation
- Self-critical review with 5 specific checks (including 'whether internal validation was overstated') prevents overconfident recommendations