From letmewatch
Extracts key frames from videos using ffmpeg scene detection, transcribes audio with optional whisper, for analyzing screen recordings, bug reports, tutorials, and demos.
npx claudepluginhub binyamineden/letmewatch --plugin letmewatchThis skill uses the workspace's default tool permissions.
Analyze video content by extracting key frames using ffmpeg scene detection and viewing them as images.
Decomposes videos into meaningful keyframes using ffmpeg scene detection filter. Extracts images on scene changes (threshold 0.01), timestamps from logs, supports MP4/MOV/WEBM/AVI/MKV. Adjusts sensitivity; warns on low frame counts indicating static videos.
Analyzes video files or YouTube URLs: extracts frames/audio, detects scenes/motion/silence/transitions via ffmpeg tools with structured workflow.
Extracts video frames with ffmpeg and reviews them via Claude's Read to verify UI flows, detect errors, stuck states, and test outcomes in videos.
Share bugs, ideas, or general feedback.
Analyze video content by extracting key frames using ffmpeg scene detection and viewing them as images.
.mp4, .mov, .mkv, .webm, .avi)/letmewatch:video, /letmewatch:video-last, or /letmewatch:video-dirbrew install ffmpeg (macOS) or apt install ffmpeg (Linux)pip install openai-whisper or pip install mlx-whisperframe_01m23s.jpg) so you can reference specific momentsRun the extraction script bundled with this skill:
python3 ${CLAUDE_PLUGIN_ROOT}/skills/letmewatch/video-extract.py "<video_path>"
Read the output to find:
WORK_DIR — where frames are storedTOTAL_FRAMES — how many frames were extractedTRANSCRIPT — path to audio transcript (or "none")FRAMES — list of frame file pathsIf TRANSCRIPT is not "none", read the transcript file first for audio/narration context.
Read frames in batches of 8 using the Read tool (all 8 in parallel). For each batch:
frame_00m23s.jpg = 0 minutes 23 seconds)After viewing all frames, provide a timestamped summary. Tailor your response:
Remove the temp directory:
rm -rf <WORK_DIR>