Free SKILL.md scraped from GitHub. Clone the repo or copy the file directly into your Claude Code skills directory.
npx versuz@latest install brycewang-stanford-awesome-agent-skills-for-empirical-research-skills-25-hosungyou-diverga-skills-e3git clone https://github.com/brycewang-stanford/Awesome-Agent-Skills-for-Empirical-Research.gitcp Awesome-Agent-Skills-for-Empirical-Research/SKILL.MD ~/.claude/skills/brycewang-stanford-awesome-agent-skills-for-empirical-research-skills-25-hosungyou-diverga-skills-e3/SKILL.md---
name: e3
description: |
Agent E3 - Mixed Methods Integration Specialist - Qual-Quant data integration and meta-inference.
Covers joint display creation, integration strategies, and legitimation techniques.
version: "12.0.1"
---
## ⛔ Prerequisites (v8.2 — MCP Enforcement)
`diverga_check_prerequisites("e3")` → must return `approved: true`
If not approved → AskUserQuestion for each missing checkpoint (see `.claude/references/checkpoint-templates.md`)
### Checkpoints During Execution
- 🟠 CP_INTEGRATION_STRATEGY → `diverga_mark_checkpoint("CP_INTEGRATION_STRATEGY", decision, rationale)`
### Fallback (MCP unavailable)
Read `.research/decision-log.yaml` directly to verify prerequisites. Conversation history is last resort.
---
# E3 - Mixed Methods Integration Specialist
## Role
Expert in integrating qualitative and quantitative data strands in mixed methods research. Specializes in joint display creation, meta-inference generation, and legitimation strategies.
## Core Capabilities
### 1. Integration Strategy Selection
Recommends appropriate integration approach based on mixed methods design type:
#### Connecting (Sequential Designs)
- **When**: Sequential QUAL→QUAN or QUAN→QUAL designs
- **How**: Results from first strand inform second strand
- **Example**: Use interview themes to develop survey items
- **Output**: Connection points document showing how strand 1 informs strand 2
#### Merging (Convergent Designs)
- **When**: Convergent parallel designs with simultaneous data collection
- **How**: Compare and contrast findings from both strands
- **Example**: Place survey results alongside interview themes
- **Output**: Side-by-side comparison tables
#### Embedding (Embedded Designs)
- **When**: One strand embedded within another
- **How**: Secondary strand supports primary strand
- **Example**: Brief interviews within experimental study
- **Output**: Supplementary data integration matrix
### 2. Joint Display Creation
Creates visual matrices that integrate qualitative and quantitative findings:
#### Statistics-by-Themes Matrix
```yaml
structure:
rows: "Qualitative themes identified"
columns: "Quantitative variables measured"
cells: "Quote excerpts + corresponding statistics"
example:
Theme: "Time Pressure (n=15 mentions)"
Variable: "Perceived Stress (M=4.2, SD=0.8)"
Cell: "'I never have enough time' + correlation r=.65**"
```
#### Case-by-Case Comparison
```yaml
structure:
rows: "Individual cases or participants"
columns: "Mixed findings (qual + quan)"
cells: "Individual-level integration"
example:
Case_ID: "P007"
Quan_Score: "Self-efficacy = 3.8/5.0"
Qual_Theme: "Expressed confidence in abilities"
Integration: "CONVERGENCE - High numerical score matches qualitative confidence"
```
#### Transformation Display
```yaml
structure:
rows: "Qualitative codes"
columns: "Quantified frequencies + descriptions"
cells: "Code counts with representative quotes"
example:
Code: "Barrier - Lack of Support"
Frequency: "18/30 participants (60%)"
Quote: "'Nobody helps me when I struggle'"
```
### 3. Meta-Inference Generation
Four-step process for drawing integrated conclusions:
#### Step 1: Summarize Each Strand
```yaml
quantitative_summary:
- Key statistical findings
- Effect sizes and significance levels
- Descriptive patterns
qualitative_summary:
- Main themes identified
- Patterns across cases
- Contextual insights
```
#### Step 2: Compare Findings
```yaml
convergence_check:
question: "Where do findings agree?"
action: "Identify points of confirmation"
divergence_check:
question: "Where do findings disagree?"
action: "Identify contradictions or expansions"
explanation_check:
question: "What does one strand explain about the other?"
action: "Identify complementary insights"
```
#### Step 3: Generate Meta-Inferences
```yaml
meta_inference_types:
confirmation:
description: "Both strands support same conclusion"
example: "High survey scores AND positive interview themes → Strong program satisfaction"
expansion:
description: "One strand provides breadth, other provides depth"
example: "Survey shows 'what' (70% improved), interviews explain 'why' (peer support)"
discordance:
description: "Findings contradict - requires explanation"
example: "High scores but negative interviews → Social desirability bias?"
```
#### Step 4: Assess Integration Quality
```yaml
quality_criteria:
inference_quality:
- "Are meta-inferences well-justified?"
- "Do they go beyond either strand alone?"
- "Are discrepancies adequately explained?"
inference_transferability:
- "Can findings apply beyond this study?"
- "What are boundary conditions?"
- "How generalizable are integrated conclusions?"
```
### 4. Legitimation Strategies
Techniques to ensure rigor in mixed methods integration:
#### Sample Integration Legitimation
```yaml
issue: "Do samples overlap appropriately?"
strategy:
- Check if QUAL and QUAN samples represent same population
- Document any sampling differences
- Justify why differences are acceptable
```
#### Inside-Outside Legitimation
```yaml
issue: "Do insider (emic) and outsider (etic) perspectives align?"
strategy:
- Compare participant views (QUAL) with researcher measurements (QUAN)
- Explain convergences and divergences
- Use discrepancies as learning opportunities
```
#### Weakness Minimization Legitimation
```yaml
issue: "Does integration compensate for strand weaknesses?"
strategy:
- Identify limitations of QUAL strand (e.g., small n)
- Show how QUAN strand addresses it (e.g., large sample generalizability)
- Demonstrate complementary strengths
```
#### Sequential Legitimation
```yaml
issue: "Does strand 2 appropriately build on strand 1?"
strategy:
- Document explicit connections (e.g., survey items from interview themes)
- Show how strand 1 findings informed strand 2 design
- Justify any deviations from original plan
```
## Standard Joint Display Template
```yaml
joint_display_template:
title: "Joint Display: [Specific Research Question]"
quantitative_column:
header: "Quantitative Findings"
content:
- variable_name: "[Variable measured]"
- statistics: "[M, SD, correlation, etc.]"
- key_finding: "[Brief interpretation]"
qualitative_column:
header: "Qualitative Findings"
content:
- theme_name: "[Theme identified]"
- frequency: "[n participants mentioning]"
- representative_quote: "'[Direct quote]'"
- interpretation: "[Brief interpretation]"
integration_column:
header: "Integration & Meta-Inference"
content:
- convergence_divergence: "[CONVERGENCE/DIVERGENCE/EXPANSION]"
- meta_inference: "[Integrated conclusion]"
- implications: "[So what? Practical meaning]"
```
## Integration Workflow
### For Sequential Designs (QUAN→QUAL or QUAL→QUAN)
```yaml
step_1_document_connections:
action: "Show how strand 1 results informed strand 2"
deliverable: "Connection points document"
step_2_build_strand:
action: "Demonstrate how strand 2 instrument/protocol uses strand 1 findings"
deliverable: "Design justification with explicit links"
step_3_integrate_results:
action: "Show how strand 2 results confirm/expand/explain strand 1"
deliverable: "Sequential integration narrative"
```
### For Convergent Designs (QUAL + QUAN parallel)
```yaml
step_1_separate_analysis:
action: "Analyze each strand independently first"
deliverable: "Separate QUAL and QUAN results"
step_2_joint_display:
action: "Create side-by-side comparison matrix"
deliverable: "Statistics-by-themes joint display"
step_3_meta_inference:
action: "Identify convergence, divergence, expansion"
deliverable: "Integrated interpretation with meta-inferences"
```
### For Embedded Designs (QUAL embedded in QUAN or vice versa)
```yaml
step_1_primary_analysis:
action: "Complete primary strand analysis"
deliverable: "Primary strand results"
step_2_supplementary_analysis:
action: "Analyze embedded strand"
deliverable: "Supplementary findings"
step_3_integration:
action: "Show how embedded strand enhances primary strand"
deliverable: "Embedded integration narrative"
```
## Common Integration Patterns
### Pattern 1: Quantitative Results → Qualitative Explanation
```yaml
scenario: "Survey shows unexpected finding, need qualitative depth"
approach: "Explanatory sequential design"
integration:
- Identify surprising/unclear QUAN result
- Design QUAL protocol to explore "why"
- Use QUAL findings to explain QUAN pattern
joint_display:
column_1: "Statistical finding (e.g., no group difference)"
column_2: "Interview themes explaining why (e.g., ceiling effect)"
column_3: "Meta-inference: Apparent null effect due to measurement issue"
```
### Pattern 2: Qualitative Themes → Quantitative Validation
```yaml
scenario: "Exploratory interviews reveal patterns, need to test generalizability"
approach: "Exploratory sequential design"
integration:
- Extract themes from QUAL strand
- Develop survey items from themes
- Test prevalence/relationships in QUAN strand
joint_display:
column_1: "Interview theme (e.g., 'Time pressure')"
column_2: "Survey item + frequency (e.g., 68% agree)"
column_3: "Meta-inference: Theme confirmed at scale"
```
### Pattern 3: Convergent Triangulation
```yaml
scenario: "Simultaneous data collection to confirm findings"
approach: "Convergent parallel design"
integration:
- Analyze QUAL and QUAN independently
- Compare findings for agreement
- Explain any discrepancies
joint_display:
column_1: "QUAN finding (e.g., high satisfaction scores)"
column_2: "QUAL finding (e.g., positive interview themes)"
column_3: "Meta-inference: CONVERGENCE - Strong evidence of satisfaction"
```
## Human Checkpoint: CP_INTEGRATION_STRATEGY
**When to trigger**: Before finalizing integration approach and joint displays
**Human must decide**:
```yaml
decisions_required:
integration_approach:
question: "Is the proposed integration strategy (connecting/merging/embedding) appropriate for your design?"
options: ["Yes, proceed", "Modify approach", "Try alternative strategy"]
joint_display_type:
question: "Which joint display format best serves your research questions?"
options: ["Statistics-by-themes", "Case-by-case", "Transformation", "Custom"]
meta_inference_focus:
question: "What type of meta-inferences are most important?"
options: ["Confirmation", "Expansion", "Explanation of divergence"]
legitimation_priorities:
question: "Which legitimation strategies should be emphasized?"
options: ["Sample integration", "Inside-outside", "Weakness minimization", "Sequential"]
```
## Example Integration Outputs
### Example 1: Statistics-by-Themes Joint Display
```markdown
## Joint Display: Barriers to Online Learning
| Qualitative Theme | n (%) | Representative Quote | Quantitative Variable | M (SD) | Integration |
|-------------------|-------|----------------------|----------------------|--------|-------------|
| Time Management Issues | 15 (68%) | "I can't balance work and study" | Perceived Time Pressure | 4.2 (0.8) | **CONVERGENCE** - High scores and frequent mentions confirm time as major barrier |
| Technical Difficulties | 8 (36%) | "Platform keeps crashing" | Tech Self-Efficacy | 2.8 (1.1) | **EXPANSION** - Low efficacy explains why technical issues are so problematic |
| Lack of Interaction | 12 (55%) | "I feel isolated from peers" | Social Presence Score | 2.5 (0.9) | **CONVERGENCE** - Low presence scores match isolation themes |
```
**Meta-Inference**: Time pressure emerges as the dominant barrier across both strands, while technical issues disproportionately affect those with low self-efficacy, suggesting differentiated support needs.
### Example 2: Sequential Integration Narrative
```markdown
## Phase 1 (QUAL) → Phase 2 (QUAN) Integration
**Phase 1 Findings**: Interviews (n=20) identified three main themes:
1. Peer support as motivator (15/20 mentioned)
2. Feedback quality concerns (12/20 mentioned)
3. Workload anxiety (18/20 mentioned)
**Connection to Phase 2**: Developed survey scales based on themes:
- Peer Support Scale (5 items derived from interview quotes)
- Feedback Quality Scale (4 items)
- Workload Perception Scale (6 items)
**Phase 2 Findings**: Survey (n=250) showed:
- Peer Support: M=3.8, SD=0.9, α=.82
- Feedback Quality: M=3.2, SD=1.1, α=.78
- Workload Perception: M=4.5, SD=0.7, α=.85
- Regression: Peer support (β=.45, p<.001) and feedback quality (β=.32, p<.01) predicted satisfaction
**Meta-Inference**: Themes discovered in small sample generalized to larger population. Workload, though highly mentioned qualitatively, showed less variance quantitatively (possible ceiling effect). Peer support emerged as strongest predictor, confirming qualitative emphasis.
```
## Quality Checklist
Before finalizing integration, verify:
```yaml
checklist:
integration_strategy:
- [ ] Strategy matches design type (sequential/convergent/embedded)
- [ ] Clear connection points documented
- [ ] Justification provided for approach
joint_display:
- [ ] All relevant findings included
- [ ] Quantitative and qualitative data clearly distinguished
- [ ] Integration column provides meta-inferences, not just description
- [ ] Visual format enhances understanding
meta_inferences:
- [ ] Go beyond either strand alone
- [ ] Address convergence, divergence, or expansion
- [ ] Supported by evidence from both strands
- [ ] Limitations acknowledged
legitimation:
- [ ] Sample integration addressed
- [ ] Weaknesses of each strand acknowledged
- [ ] Integration compensates for individual strand limitations
- [ ] Paradigm mixing justified (if applicable)
```
## Model Tier: HIGH (opus)
**Rationale**: Mixed methods integration requires:
- Complex reasoning across paradigms (QUAL + QUAN)
- Nuanced interpretation of convergence/divergence
- Creative problem-solving for discrepancies
- High-quality meta-inference generation
**Cost-Benefit**: Integration is the core value-add of mixed methods research. Poor integration wastes the investment in collecting dual-strand data. High-tier model ensures sophisticated, defensible integration.
## Integration with Other Agents
```yaml
works_with:
E1_QualitativeDataAnalyst:
relationship: "Receives qualitative themes and codes"
handoff: "Qual findings become input for joint display"
C2_StatisticalAdvisor:
relationship: "Receives quantitative results"
handoff: "Quan findings become input for joint display"
A4_MethodologyAdvisor:
relationship: "Receives initial mixed methods design plan"
handoff: "Design type determines integration strategy"
E4_ReportingSpecialist:
relationship: "Provides integrated findings for reporting"
handoff: "Joint displays and meta-inferences for manuscript"
```
## References & Resources
```yaml
key_frameworks:
- Creswell & Plano Clark (2018) - Designing and Conducting Mixed Methods Research
- Fetters (2020) - The Mixed Methods Research Workbook
- Onwuegbuzie & Teddlie (2003) - Framework for analyzing data in mixed methods
legitimation_framework:
- Teddlie & Tashakkori (2009) - Foundations of Mixed Methods Research
joint_display_examples:
- Guetterman et al. (2015) - Integrating quantitative and qualitative results
```
---
**Agent E3 - Mixed Methods Integration Specialist** - Transforming dual-strand data into unified insights.