faster-whisperLocal speech-to-text using faster-whisper. 4-6x faster than OpenAI Whisper with identical accuracy; GPU acceleration enables ~20x realtime transcription. SRT...
Install via ClawdBot CLI:
clawdbot install theplasmak/faster-whisperGrade Good — based on market validation, documentation quality, package completeness, maintenance status, and authenticity signals.
Accesses system directories or attempts privilege escalation
/proc/Calls external URL not in known-safe list
https://github.com/ThePlasmak/faster-whisperUses known external API (expected, informational)
arxiv.orgAI Analysis
The skill's primary function is local transcription using a known open-source library, and its external calls (GitHub, arXiv) are for model downloads and documentation, consistent with its purpose. The flagged system directory access is likely for hardware detection (GPU/CUDA) or process management, not privilege escalation, and no evidence suggests hidden instructions, credential harvesting, or data exfiltration.
Generated Mar 1, 2026
Podcasters can transcribe episodes for SEO, accessibility, and show notes. The skill supports batch processing of RSS feeds, speaker diarization to label hosts and guests, and export of subtitles in formats like SRT for video platforms.
Educational institutions can transcribe lectures and seminars for student notes, closed captions, and archival. The skill handles multilingual content, detects chapters for easy navigation, and allows search within transcripts for study purposes.
Businesses can transcribe meetings, interviews, and conferences to create searchable records and action items. Features like speaker diarization identify participants, while CSV output facilitates analysis and reporting in spreadsheets.
Content creators and media companies can generate subtitles in multiple formats (e.g., SRT, VTT) for videos, including YouTube links. The skill supports translation to English and batch processing, enabling efficient localization for global audiences.
Healthcare providers can transcribe patient consultations and medical notes locally for privacy. The skill's initial-prompt feature helps with medical jargon, while denoising improves accuracy in noisy clinical environments.
Offer a cloud-based platform where users upload audio files for automated transcription. Leverage the skill's GPU acceleration for fast processing and charge per minute of audio, with premium features like speaker diarization and custom formats.
Sell licensed software packages to large organizations needing offline, secure transcription. Bundle with support and training, using the skill's batch processing and integration capabilities for internal workflows like legal or media production.
Operate a service where freelancers use the tool to transcribe audio for clients in industries like academia or podcasting. Scale by automating bulk jobs with batch processing and offering add-ons like subtitle formatting or translation.
💬 Integration Tip
Ensure Python3 and optional tools like ffmpeg are installed; cache Hugging Face tokens for seamless model access to avoid interruptions.
Scored Apr 16, 2026
Audited Apr 16, 2026 · audit v1.0
Transcribe audio via OpenAI Audio Transcriptions API (Whisper).
Local speech-to-text with the Whisper CLI (no API key).
ElevenLabs text-to-speech with mac-style say UX.
Local text-to-speech via sherpa-onnx (offline, no cloud)
Speak responses aloud on macOS using the built-in `say` command when user input indicates Voice Wake/voice recognition (for example, messages starting with "User talked via voice recognition on <device>").
Transcribe audio files to text using local Whisper (Docker). Use when receiving voice messages, audio files (.mp3, .m4a, .ogg, .wav, .webm), or when asked to transcribe audio content.