Free SKILL.md scraped from GitHub. Clone the repo or copy the file directly into your Claude Code skills directory.
npx versuz@latest install hiyenwong-ai-collection-collection-skills-eeg2vision-multimodal-eeg-framework-2dgit clone https://github.com/hiyenwong/ai_collection.gitcp ai_collection/SKILL.MD ~/.claude/skills/hiyenwong-ai-collection-collection-skills-eeg2vision-multimodal-eeg-framework-2d/SKILL.md---
name: eeg2vision-multimodal-eeg-framework-2d
description: "EEG2Vision: A multimodal EEG-based framework for 2D visual reconstruction from non-invasive brain signals. Combines EEG encoding with diffusion models for image generation. Activation: eeg, visual reconstruction, diffusion model, brain imaging, visual decoding, multimodal, eeg2image, brain-to-image"
---
# EEG2Vision: Multimodal EEG-Based Framework for 2D Visual Reconstruction
## Overview
A modular framework that reconstructs visual stimuli from non-invasive EEG signals. The system bridges the gap between low-resolution EEG data and high-fidelity visual reconstruction by leveraging multimodal learning and diffusion-based generation.
## Source Paper
- **Title:** EEG2Vision: A Multimodal EEG-Based Framework for 2D Visual Reconstruction
- **arXiv:** Available on arXiv
- **Categories:** neuroscience, EEG, computer vision
## Core Concepts
### Architecture
1. **EEG Encoder**: Extract spatiotemporal features from raw EEG signals
2. **Visual Latent Mapper**: Map EEG features to visual latent space
3. **Diffusion Generator**: Generate images conditioned on EEG-derived latents
4. **Multimodal Alignment**: Align EEG and visual representations
### Key Innovations
- **Non-invasive visual decoding**: Works with EEG (not fMRI/iEEG)
- **Diffusion-based generation**: Leverages pre-trained image diffusion models
- **Cross-modal alignment**: Learn shared representation space
- **Temporal feature extraction**: Capture time-varying visual processing in EEG
## Implementation Pattern
```python
class EEG2VisionPipeline:
"""End-to-end EEG to visual reconstruction pipeline."""
def __init__(self, eeg_channels=64, latent_dim=512):
# EEG feature extractor
self.eeg_encoder = EEGTemporalEncoder(
channels=eeg_channels,
num_layers=4,
hidden_dim=256
)
# Cross-modal mapper
self.latent_mapper = CrossModalMapper(
input_dim=256,
output_dim=latent_dim
)
# Image generation (frozen diffusion model)
self.diffusion_model = StableDiffusion(latent_dim)
def train(self, eeg_data, images, num_epochs=100):
"""Train EEG encoder and latent mapper."""
for epoch in range(num_epochs):
# Extract EEG features
eeg_features = self.eeg_encoder(eeg_data)
# Map to visual latent space
latents = self.latent_mapper(eeg_features)
# Compute alignment loss
image_features = self.get_image_features(images)
alignment_loss = self.contrastive_loss(latents, image_features)
# Reconstruction loss via diffusion
recon_loss = self.diffusion_model.compute_loss(latents, images)
total_loss = alignment_loss + recon_loss
self.backward(total_loss)
def reconstruct(self, eeg_signal):
"""Reconstruct image from EEG."""
features = self.eeg_encoder(eeg_signal)
latents = self.latent_mapper(features)
image = self.diffusion_model.generate(latents)
return image
```
## Practical Applications
### Clinical
- Visual prosthesis research
- Understanding visual processing disorders
- Non-invasive visual communication aids
### Research
- Study of visual perception from EEG
- Cross-modal representation learning
- Brain-computer interface development
## Limitations
- EEG has lower spatial resolution than fMRI/iEEG
- Reconstruction quality depends on training data diversity
- Limited to visual categories seen during training
- Temporal resolution mismatch between EEG and visual processing
## Activation Keywords
- eeg
- visual reconstruction
- diffusion model
- brain imaging
- visual decoding
- multimodal
- eeg2image