From sundial-org-awesome-openclaw-skills-4
Transcribes audio to text locally using MLX Whisper on Apple Silicon Macs. Supports text output, SRT subtitles, translation, language hints without API keys.
npx claudepluginhub joshuarweaver/cascade-ai-ml-agents-misc-2 --plugin sundial-org-awesome-openclaw-skills-4This skill uses the workspace's default tool permissions.
Local speech-to-text using Apple MLX, optimized for Apple Silicon Macs.
Guides Next.js Cache Components and Partial Prerendering (PPR) with cacheComponents enabled. Implements 'use cache', cacheLife(), cacheTag(), revalidateTag(), static/dynamic optimization, and cache debugging.
Guides building MCP servers enabling LLMs to interact with external services via tools. Covers best practices, TypeScript/Node (MCP SDK), Python (FastMCP).
Generates original PNG/PDF visual art via design philosophy manifestos for posters, graphics, and static designs on user request.
Local speech-to-text using Apple MLX, optimized for Apple Silicon Macs.
mlx_whisper /path/to/audio.mp3 --model mlx-community/whisper-large-v3-turbo
# Transcribe to text file
mlx_whisper audio.m4a -f txt -o ./output
# Transcribe with language hint
mlx_whisper audio.mp3 --language en --model mlx-community/whisper-large-v3-turbo
# Generate subtitles (SRT)
mlx_whisper video.mp4 -f srt -o ./subs
# Translate to English
mlx_whisper foreign.mp3 --task translate
| Model | Size | Speed | Quality |
|---|---|---|---|
| mlx-community/whisper-tiny | ~75MB | Fastest | Basic |
| mlx-community/whisper-base | ~140MB | Fast | Good |
| mlx-community/whisper-small | ~470MB | Medium | Better |
| mlx-community/whisper-medium | ~1.5GB | Slower | Great |
| mlx-community/whisper-large-v3 | ~3GB | Slowest | Best |
| mlx-community/whisper-large-v3-turbo | ~1.6GB | Fast | Excellent (Recommended) |
~/.cache/huggingface/mlx-community/whisper-tiny; use --model mlx-community/whisper-large-v3-turbo for best results