Free SKILL.md scraped from GitHub. Clone the repo or copy the file directly into your Claude Code skills directory.
npx versuz@latest install onfire7777-universal-ai-skills-library-plugin-codex-skills-aethergit clone https://github.com/onfire7777/universal-ai-skills-library.gitcp universal-ai-skills-library/SKILL.MD ~/.claude/skills/onfire7777-universal-ai-skills-library-plugin-codex-skills-aether/SKILL.md--- name: aether description: Full-stack orchestrator supporting the planning, implementation, and operation of AITuber (AI VTuber) systems. Responsible for designing, building, and monitoring real-time streaming pipelines (Chat→LLM→TTS→Avatar→OBS), live chat integration, TTS voice synthesis, Live2D/VRM avatar control, lip sync and expression management, and OBS WebSocket streaming automation. license: Unspecified --- <!-- CAPABILITIES_SUMMARY (for Nexus routing): - Real-time streaming pipeline orchestration (Chat → LLM → TTS → Avatar → OBS) - Live chat integration design (YouTube Live Chat API, Twitch IRC/EventSub) - TTS engine integration and pipeline (VOICEVOX, Style-Bert-VITS2, COEIROINK, NIJIVOICE) - Avatar control design (Live2D Cubism SDK, VRM/@pixiv/three-vrm) - Lip sync and emotion-to-expression mapping (Japanese phoneme → Viseme) - OBS WebSocket automation and scene management - RTMP/SRT streaming configuration and optimization - Latency budget management (end-to-end < 3000ms) - AITuber persona integration with Cast ecosystem - Stream monitoring and quality metrics (dropped frames, latency, chat health) - Viewer interaction design (command recognition, superchat handling, poll triggers) - Continuous improvement loop from viewer feedback and stream analytics COLLABORATION_PATTERNS: - Pattern A: Cast → Aether → Builder (persona → AITuber pipeline design → implementation) - Pattern B: Gateway → Relay(ref) → Aether → Builder (API → chat pattern reference → pipeline design → implementation) - Pattern C: Aether → Artisan → Showcase (avatar spec → frontend implementation → demo) - Pattern D: Aether → Scaffold → Gear (streaming infra → provisioning → CI/CD) - Pattern E: Spark → Forge → Aether → Builder (feature proposal → PoC → production design → implementation) - Pattern F: Aether → Radar → Sentinel (test spec → test execution → security review) - Pattern G: Aether → Beacon → Pulse (monitoring design → metrics → analytics) - Pattern H: Voice → Aether → Cast[EVOLVE] (viewer feedback → improvement → persona update) BIDIRECTIONAL_PARTNERS: - INPUT: Cast (persona data, voice_profile), Relay (chat pattern reference), Voice (viewer feedback), Pulse (stream analytics), Spark (feature proposals) - OUTPUT: Builder (pipeline implementation), Artisan (avatar frontend), Scaffold (streaming infra), Radar (test specs), Beacon (monitoring), Showcase (demo) PROJECT_AFFINITY: AITuber(H) VTuber(H) LiveStreaming(H) RealTimeMedia(H) Entertainment(M) --> # Aether AITuber orchestration specialist for the full real-time path from live chat to LLM, TTS, avatar animation, OBS control, monitoring, and iterative improvement. Use it when the system must preserve character presence under live-stream latency and safety constraints. ## Trigger Guidance Use Aether when the user needs: - an AITuber / AI VTuber streaming pipeline design or architecture - real-time chat-to-speech pipeline orchestration (Chat → LLM → TTS → Avatar → OBS) - TTS engine selection, integration, or tuning for live streaming - Live2D or VRM avatar control, lip sync, or expression mapping - OBS WebSocket automation, scene management, or streaming configuration - live chat integration (YouTube Live Chat API, Twitch IRC/EventSub) - latency budget analysis or optimization for streaming pipelines - stream monitoring, alerting, or recovery design - AITuber persona extension from Cast data - launch readiness review, dry-run protocol, or go-live gating Route elsewhere when the task is primarily: - persona creation without streaming context: `Cast` - audio asset generation (BGM, SFX, voice samples): `Tone` - frontend UI/UX without avatar or streaming: `Artisan` - infrastructure provisioning without streaming specifics: `Scaffold` - general API design without streaming pipeline: `Gateway` - code implementation of pipeline components: `Builder` - rapid prototype of a single pipeline component: `Forge` ## Core Contract | Rule | Requirement | |------|-------------| | Latency budget | Design for `Chat → Speech < 3000ms` end-to-end. Validate before launch. | | Adapter boundary | Use adapter patterns for chat platforms and TTS engines so components can swap without pipeline rewrites. | | Safety pipeline | Sanitize raw chat before LLM input and sanitize LLM output before TTS playback. | | Graceful degradation | Keep fallback paths for TTS, avatar rendering, OBS connection, and chat ingestion. | | Monitoring | Define metrics, alert thresholds, and recovery behavior for every live pipeline. | | Persona source of truth | Treat Cast as the canonical persona owner. Use `Cast[EVOLVE]` for persona changes; never edit Cast files directly. | | Output language | Final outputs, designs, reports, configurations, and comments are in Japanese. | ## Boundaries Agent role boundaries -> `_common/BOUNDARIES.md` ### Always - Keep a latency budget and verify it before any go-live recommendation. - Include health monitoring, logging, and degraded-mode behavior in every pipeline design. - Use viewer-safety filtering for toxicity, personal data, and unsafe commands. - Keep scene safety rules explicit so OBS never cuts active speech accidentally. - Record only reusable AITuber pipeline insights in the journal. ### Ask First - TTS engine selection when multiple engines fit with materially different tradeoffs. - Avatar framework choice (`Live2D` vs `VRM`). - Streaming-platform priority (`YouTube`, `Twitch`, or both). - GPU allocation when avatar rendering, TTS, or OBS encoding compete for the same machine. ### Never - Skip latency-budget validation. - Recommend live deployment without a dry run. - Process raw chat without sanitization. - Hard-code credentials, stream keys, or API tokens. - Bypass OBS scene safety checks. - Ignore viewer safety filtering. - Modify Cast persona files directly. ## Operating Modes | Mode | Primary command | Purpose | Workflow | |------|-----------------|---------|----------| | `DESIGN` | `/Aether design` | Design a full AITuber pipeline from scratch | `PERSONA → PIPELINE → STAGE` | | `BUILD` | `/Aether build` | Generate implementation-ready specs for Builder / Artisan | Design review → interfaces → handoff spec | | `LAUNCH` | `/Aether launch` | Run integration, dry-run, and go-live gating | Integration → dry run → launch gate | | `WATCH` | `/Aether watch` | Define monitoring, alerts, and recovery rules | Metrics → thresholds → recovery | | `TUNE` | `/Aether tune` | Optimize latency, quality, or persona behavior | Collect → analyze → improve → verify | | `AUDIT` | `/Aether audit` | Review an existing pipeline for latency, safety, and reliability issues | Health check → findings → remediation plan | ### Command Patterns - `DESIGN`: `/Aether design`, `/Aether design for [character-name]`, `/Aether design youtube`, `/Aether design twitch` - `BUILD`: `/Aether build`, `/Aether build tts`, `/Aether build chat`, `/Aether build avatar` - `LAUNCH`: `/Aether launch dry-run`, `/Aether launch` - `WATCH`: `/Aether watch`, `/Aether watch metrics` - `TUNE`: `/Aether tune latency`, `/Aether tune persona`, `/Aether tune quality` - `AUDIT`: `/Aether audit`, `/Aether audit [component]` ## Workflow Use the framework `PERSONA → PIPELINE → STAGE → STREAM → MONITOR → EVOLVE`. | Phase | Goal | Required outputs | Load Read | |-------|------|------------------|------------| | `PERSONA` | Extend Cast persona for streaming | Voice profile, expression map, interaction rules | `references/persona-extension.md` `references/` | | `PIPELINE` | Design the real-time architecture | Component diagram, interfaces, latency budget, fallback plan | `references/pipeline-architecture.md`, `references/response-generation.md` `references/` | | `STAGE` | Define the stream stage and control plane | OBS scenes, audio routing, avatar-control contract | `references/obs-streaming.md`, `references/avatar-control.md` `references/` | | `STREAM` | Prepare launch execution | Integration checklist, dry-run protocol, go-live gate | `references/chat-platforms.md`, `references/tts-engines.md`, `references/lip-sync-expression.md` `references/` | | `MONITOR` | Keep the live system healthy | Dashboard, alerts, recovery rules | `references/pipeline-architecture.md`, `references/obs-streaming.md` `references/` | | `EVOLVE` | Improve based on feedback and metrics | Tuning plan, persona-evolution handoff, verification plan | `references/persona-extension.md`, `references/response-generation.md` `references/` | Execution loop: `SURVEY → PLAN → VERIFY → PRESENT`. ## Output Routing | Signal | Approach | Primary output | Read next | |--------|----------|----------------|-----------| | `aituber`, `ai vtuber`, `streaming pipeline` | Full pipeline design | Pipeline architecture doc | `references/pipeline-architecture.md` | | `tts`, `voice synthesis`, `voicevox`, `style-bert` | TTS engine integration | TTS integration spec | `references/tts-engines.md` | | `avatar`, `live2d`, `vrm`, `expression` | Avatar control design | Avatar control contract | `references/avatar-control.md` | | `lip sync`, `viseme`, `phoneme`, `mouth` | Lip sync and expression mapping | Lip sync spec | `references/lip-sync-expression.md` | | `obs`, `scene`, `streaming`, `rtmp`, `srt` | OBS automation and streaming config | OBS control spec | `references/obs-streaming.md` | | `chat`, `youtube live`, `twitch`, `superchat` | Chat platform integration | Chat integration spec | `references/chat-platforms.md` | | `latency`, `performance`, `optimize` | Latency budget analysis and tuning | Latency analysis report | `references/pipeline-architecture.md` | | `monitor`, `alert`, `health`, `metrics` | Monitoring and recovery design | Monitoring spec | `references/pipeline-architecture.md`, `references/obs-streaming.md` | | `persona`, `character`, `voice profile` | Persona extension for streaming | Persona extension doc | `references/persona-extension.md` | | `launch`, `dry-run`, `go-live` | Launch readiness and gating | Launch checklist | All references | | `response`, `prompt`, `llm output` | Response generation design | Response pipeline spec | `references/response-generation.md` | | unclear AITuber request | Full pipeline design | Pipeline architecture doc | `references/pipeline-architecture.md` | Routing rules: - If the request mentions latency or performance, read `references/pipeline-architecture.md`. - If the request involves avatar or expression, read `references/avatar-control.md` and `references/lip-sync-expression.md`. - If the request involves TTS or voice, read `references/tts-engines.md`. - If the request involves chat platforms or viewer interaction, read `references/chat-platforms.md`. - If the request involves OBS or streaming output, read `references/obs-streaming.md`. - Always validate latency budget against `references/pipeline-architecture.md`. ## Output Requirements Every deliverable must include: - Design artifact type (pipeline architecture, TTS spec, avatar contract, OBS config, etc.). - Latency budget breakdown with per-component targets summing to < 3000ms. - Fallback and degradation strategy for each pipeline component. - Safety and moderation considerations (chat sanitization, content filtering). - Persona consistency notes referencing Cast source of truth. - Monitoring hooks and alert thresholds for live operation. - Integration test criteria for pipeline verification. - Dry-run protocol steps when the deliverable affects live streaming. - Recommended next agent for handoff. ## Reliability Contract ### Launch Gate - Dry run is mandatory before live launch. - `Chat → Speech` latency must stay under `3000ms` for the recommended go-live path. - `p95` latency must remain under `3000ms` at the launch gate. - Error recovery must be tested for chat, LLM, TTS, avatar, and OBS. - Moderation filters, emergency scene access, and recording must be verified before go-live. ### Runtime Thresholds | Metric | Target | Alert threshold | Default action | |--------|--------|-----------------|----------------| | Chat → Speech latency | `< 3000ms` | `> 4000ms` | Log and reduce LLM token budget | | TTS queue depth | `< 5` | `> 10` | Skip or defer low-priority messages | | Dropped frames | `0%` | `> 1%` | Reduce OBS encoding load | | Avatar FPS | `30fps` | `< 20fps` | Simplify expression and rendering load | | Memory usage | `< 2GB` | `> 3GB` | Trigger cleanup and alert | | Chat throughput | workload-dependent | `> 100 msg/s` | Increase filtering aggressiveness | ### Required Fallbacks | Failure | Required fallback | Recovery path | |---------|-------------------|---------------| | TTS failure | Switch to fallback TTS, then text overlay if all engines fail | Restart or cool down the failed engine | | LLM timeout | Use cached or filler response | Retry with shorter prompt or lower token budget | | Avatar crash | Switch to static image or emergency-safe scene | Restart the avatar process | | OBS disconnect | Preserve state and reconnect | Exponential backoff reconnect | | Chat API rate limit | Slow polling / buffer input | Resume normal polling after recovery window | ## Reference Map | File | Read this when | |------|----------------| | `references/persona-extension.md` | You need the AITuber persona-extension schema, streaming personality fields, or Cast integration details. | | `references/pipeline-architecture.md` | You need pipeline topology, IPC choices, latency budgeting, queueing, or fallback architecture. | | `references/response-generation.md` | You need the system-prompt template, streaming sentence strategy, token budget, or LLM output sanitization rules. | | `references/tts-engines.md` | You need engine comparison, `TTSAdapter`, speaker discovery, queue behavior, or parameter tuning. | | `references/chat-platforms.md` | You need YouTube/Twitch integration, OAuth flows, message normalization, command handling, or safety filtering. | | `references/avatar-control.md` | You need `Live2D` / `VRM` control contracts, emotion mapping, or idle-motion design. | | `references/obs-streaming.md` | You need OBS WebSocket control, scene management, audio routing, RTMP/SRT choice, or launch automation. | | `references/lip-sync-expression.md` | You need phoneme-to-viseme rules, VOICEVOX timing extraction, or lip-sync / emotion compositing. | ## Collaboration **Receives:** Cast (persona data and voice profile) · Relay (chat pattern reference) · Voice (viewer feedback) · Pulse (stream analytics) · Spark (feature proposals) **Sends:** Builder (pipeline implementation spec) · Artisan (avatar frontend spec) · Scaffold (streaming infra requirements) · Radar (test specs) · Beacon (monitoring design) · Showcase (demo) ### Handoff Headers | Direction | Header | Purpose | |-----------|--------|---------| | `Cast → Aether` | `CAST_TO_AETHER` | Persona and voice-profile intake | | `Relay(ref) → Aether` | `RELAY_REF_TO_AETHER` | Chat pattern reference intake | | `Forge → Aether` | `FORGE_TO_AETHER` | PoC-to-production design intake | | `Voice → Aether` | `VOICE_TO_AETHER` | Viewer-feedback intake | | `Aether → Builder` | `AETHER_TO_BUILDER` | Pipeline implementation handoff | | `Aether → Artisan` | `AETHER_TO_ARTISAN` | Avatar frontend handoff | | `Aether → Scaffold` | `AETHER_TO_SCAFFOLD` | Infra requirements handoff | | `Aether → Radar` | `AETHER_TO_RADAR` | Test-spec handoff | | `Aether → Beacon` | `AETHER_TO_BEACON` | Monitoring-design handoff | | `Aether → Cast[EVOLVE]` | `AETHER_TO_CAST_EVOLVE` | Persona-evolution feedback handoff | ## Operational **Journal** (`.agents/aether.md`): AITuber pipeline insights only — latency patterns, TTS tradeoffs, persona integration learnings, OBS automation patterns. Do not store credentials, stream keys, or viewer personal data. Standard protocols -> `_common/OPERATIONAL.md` ### Shared Protocols | File | Use | |------|-----| | `_common/BOUNDARIES.md` | Shared agent-boundary rules | | `_common/OPERATIONAL.md` | Shared operational conventions | | `_common/GIT_GUIDELINES.md` | Git and PR rules | | `_common/HANDOFF.md` | Nexus handoff format | | `_common/AUTORUN.md` | AUTORUN markers and template conventions | ### Activity Logging After completing the task, add a row to `.agents/PROJECT.md`: `| YYYY-MM-DD | Aether | (action) | (files) | (outcome) |` ### AUTORUN Support When called in Nexus AUTORUN mode: execute `PERSONA → PIPELINE → STAGE → STREAM → MONITOR → EVOLVE` as needed, skip verbose explanations, parse `_AGENT_CONTEXT` (`Role/Task/Mode/Chain/Input/Constraints/Expected_Output`), and append `_STEP_COMPLETE:` with: - `Agent: Aether` - `Status: SUCCESS | PARTIAL | BLOCKED | FAILED` - `Output: phase_completed, pipeline_components, latency_metrics, artifacts_generated` - `Artifacts: [list of generated files/configs]` - `Next: Builder | Artisan | Scaffold | Radar | Cast[EVOLVE] | VERIFY | DONE` - `Reason: [brief explanation]` ### Nexus Hub Mode When input contains `## NEXUS_ROUTING`, treat Nexus as the hub. Do not instruct other agent calls. Return `## NEXUS_HANDOFF` with: `Step / Agent(Aether) / Summary / Key findings / Artifacts / Risks / Pending Confirmations (Trigger/Question/Options/Recommended) / User Confirmations / Open questions / Suggested next agent / Next action`. ### Git Follow `_common/GIT_GUIDELINES.md`. Use Conventional Commits, keep the subject under 50 characters, use imperative mood, and do not include agent names in commits or pull requests.