From sundial-org-awesome-openclaw-skills-4
Uses macOS `say` command to speak responses aloud when messages start with 'User talked via voice recognition', providing voice feedback for voice wake inputs.
npx claudepluginhub joshuarweaver/cascade-ai-ml-agents-misc-2 --plugin sundial-org-awesome-openclaw-skills-4This skill uses the workspace's default tool permissions.
Use macOS `say` to read the assistant's response out loud whenever the conversation came from Voice Wake/voice recognition. Do **not** use the `tts` tool (it calls cloud providers).
Guides Next.js Cache Components and Partial Prerendering (PPR) with cacheComponents enabled. Implements 'use cache', cacheLife(), cacheTag(), revalidateTag(), static/dynamic optimization, and cache debugging.
Guides building MCP servers enabling LLMs to interact with external services via tools. Covers best practices, TypeScript/Node (MCP SDK), Python (FastMCP).
Generates original PNG/PDF visual art via design philosophy manifestos for posters, graphics, and static designs on user request.
Use macOS say to read the assistant's response out loud whenever the conversation came from Voice Wake/voice recognition. Do not use the tts tool (it calls cloud providers).
say (CHECK EVERY MESSAGE INDIVIDUALLY)IF the user message STARTS WITH: User talked via voice recognition
say first (so the user knows you heard them)IF the user message does NOT start with that exact phrase
say. Text-only response only.Critical:
User talked via voice recognitionsay (local macOS TTS)printf '%s' "$SPOKEN_TEXT" | say
Optional controls (use only if set):
printf '%s' "$SPOKEN_TEXT" | say -v "$SAY_VOICE"
printf '%s' "$SPOKEN_TEXT" | say -r "$SAY_RATE"
say is unavailable or errors, still send the text response and note that TTS failed.