Jump to related tools in the same category or review the original source on GitHub.

Apple Apps & Services @xadenryan Updated 1/25/2026

Clawdbot Skill Voice Wake Say OpenClaw Plugin & Skill | ClawHub

Looking to integrate Clawdbot Skill Voice Wake Say into your AI workflows? This free OpenClaw plugin from ClawHub helps you automate apple apps & services tasks instantly, without having to write custom tools from scratch.

What this skill does

Speak responses aloud on macOS using the built-in `say` command when user input indicates Voice Wake/voice recognition (for example, messages starting with "User talked via voice recognition on <device>").

Install

npx clawhub@latest install clawdbot-skill-voice-wake-say

Full SKILL.md

Open original
Metadata table.
namedescription
voice-wake-saySpeak responses aloud on macOS using the built-in `say` command when user input indicates Voice Wake/voice recognition (for example, messages starting with "User talked via voice recognition on <device>").

SKILL.md content below is scrollable.

Voice Wake Say

Overview

Use macOS say to read the assistant's response out loud whenever the conversation came from Voice Wake/voice recognition. Do not use the tts tool (it calls cloud providers).

When to Use say (CHECK EVERY MESSAGE INDIVIDUALLY)

IF the user message STARTS WITH: User talked via voice recognition on m3

  • Step 1: Acknowledge with say first (so the user knows you heard them)
  • Step 2: Then perform the task
  • Step 3: Optionally speak again when done if it makes sense

IF the user message does NOT start with that exact phrase

  • THEN: Do NOT use say. Text-only response only.

Critical:

  • Check EACH message individually — context does NOT carry over
  • The trigger phrase must be at the VERY START of the message
  • For tasks that take time, acknowledge FIRST so the user knows you're working

Workflow

  1. Detect Voice Wake context
  • Trigger ONLY when the latest user/system message STARTS WITH User talked via voice recognition on m3
  • If the message instructs "repeat prompt first", keep that behavior in the response.
  1. Prepare spoken text
  • Use the final response text as the basis.
  • Strip markdown/code blocks; if the response is long or code-heavy, speak a short summary and mention that details are on screen.
  1. Speak with say (local macOS TTS)
printf '%s' "$SPOKEN_TEXT" | say

Optional controls (use only if set):

printf '%s' "$SPOKEN_TEXT" | say -v "$SAY_VOICE"
printf '%s' "$SPOKEN_TEXT" | say -r "$SAY_RATE"

Failure handling

  • If say is unavailable or errors, still send the text response and note that TTS failed.
Original Repository URL: https://github.com/openclaw/skills/blob/main/skills/xadenryan/clawdbot-skill-voice-wake-say
Latest commit: https://github.com/openclaw/skills/commit/1edddd6701d3f91d9faedf40f8d67bab94986d85

Related skills

If this matches your use case, these are close alternatives in the same category.