From antigravity-awesome-skills
Guides design, implementation, and optimization of CV pipelines with YOLO26 detection, SAM 3 segmentation, VLMs, and spatial analysis for real-time systems.
npx claudepluginhub mit-network/antigravity-awesome-skillsThis skill uses the workspace's default tool permissions.
**Role**: Advanced Vision Systems Architect & Spatial Intelligence Expert
Provides expert guidance on YOLO26 detection, SAM 3 segmentation, VLMs, depth estimation, and 3D reconstruction for real-time computer vision pipelines.
Guides implementation of object detection (YOLO), semantic/instance segmentation (SAM), 3D vision, depth estimation, video understanding, and multi-modal vision using reference patterns, diagnostics, and validations.
Provides expert engineering for VLM segmentation pipelines (SAM3, Grounding DINO, YOLO-World), diffusion models (UNet, DiT, Flux, LoRA), and GPU deployment (MIG, Triton, H100 optimization).
Share bugs, ideas, or general feedback.
Role: Advanced Vision Systems Architect & Spatial Intelligence Expert
To provide expert guidance on designing, implementing, and optimizing state-of-the-art computer vision pipelines. From real-time object detection with YOLO26 to foundation model-based segmentation with SAM 3 and visual reasoning with VLMs.
| Issue | Severity | Solution |
|---|---|---|
| SAM 3 VRAM Usage | Medium | Use quantized/distilled versions for local GPU inference. |
| Text Ambiguity | Low | Use descriptive prompts ("the 5mm bolt" instead of just "bolt"). |
| Motion Blur | Medium | Optimize shutter speed or use SAM 3's temporal tracking consistency. |
| Hardware Compatibility | Low | YOLO26 simplified architecture is highly compatible with NPU/TPUs. |
ai-engineer, robotics-expert, research-engineer, embedded-systems