dialogue-audioMulti-speaker dialogue audio creation with Dia TTS. Covers speaker tags, emotion control, pacing, conversation flow, and post-production. Use for: podcasts,...
Install via ClawdBot CLI:
clawdbot install okaris/dialogue-audioGrade Fair — based on market validation, documentation quality, package completeness, maintenance status, and authenticity signals.
Calls external URL not in known-safe list
https://inference.shAudited Apr 16, 2026 · audit v1.0
Generated Mar 20, 2026
Enables creators to generate dialogue segments for podcasts without needing multiple voice actors. Useful for scripted shows, narrative podcasts, or filler content, saving time and cost on recording sessions.
Facilitates multi-character audiobook creation by assigning distinct voices to different speakers in dialogues. Ideal for indie authors or publishers looking to produce audio versions efficiently with consistent voice quality.
Supports the creation of conversational audio for tutorials, online courses, or instructional content. Helps simulate teacher-student interactions or character-based explanations to enhance engagement in e-learning materials.
Allows game developers to prototype or produce dialogue for non-player characters (NPCs) with emotional cues and pacing control. Useful for indie studios needing quick voiceovers during development phases.
Generates realistic dialogue for training scenarios, such as customer service simulations or role-playing exercises. Helps organizations create interactive audio content without hiring actors for each session.
Offer a platform where users pay a monthly fee to access advanced dialogue generation features, such as custom voice options or higher usage limits. Targets content creators and businesses needing regular audio production.
Provide an API that charges per audio minute generated, allowing developers and companies to integrate dialogue creation into their apps or workflows. Scales with usage and appeals to tech-savvy users.
Offer a free tier with basic features like two-speaker dialogues and limited emotion control, while premium tiers unlock advanced options like more speakers or post-production tools. Attracts hobbyists and upsells to professionals.
💬 Integration Tip
Integrate via the inference.sh CLI with simple JSON inputs; ensure speaker tags and punctuation are correctly formatted for optimal audio output.
Scored Apr 19, 2026
Local speech-to-text with the Whisper CLI (no API key).
ElevenLabs text-to-speech with mac-style say UX.
Transcribe audio via OpenAI Audio Transcriptions API (Whisper).
Text-to-speech conversion using node-edge-tts npm package for generating audio from text. Supports multiple voices, languages, speed adjustment, pitch control, and subtitle generation. Use when: (1) User requests audio/voice output with the "tts" trigger or keyword. (2) Content needs to be spoken rather than read (multitasking, accessibility, driving, cooking). (3) User wants a specific voice, speed, pitch, or format for TTS output.
Local text-to-speech via sherpa-onnx (offline, no cloud)
Start voice calls via the OpenClaw voice-call plugin.