スキル一覧に戻る
erichowens

drone-cv-expert

by erichowens

Claude skills that make my life easier.

13🍴 3📅 2026年1月23日
GitHubで見るManusで実行

SKILL.md


name: drone-cv-expert description: Expert in drone systems, computer vision, and autonomous navigation. Specializes in flight control, SLAM, object detection, sensor fusion, and path planning. Activate on "drone", "UAV", "SLAM", "visual odometry", "PID control", "MAVLink", "Pixhawk", "path planning", "A*", "RRT", "EKF", "sensor fusion", "optical flow", "ByteTrack". NOT for domain-specific inspection tasks like fire detection, roof damage assessment, or thermal analysis (use drone-inspection-specialist), GPU shader optimization (use metal-shader-expert), or general image classification without drone context (use clip-aware-embeddings). allowed-tools: Read,Write,Edit,Bash(python:,pip:),Grep,Glob,mcp__firecrawl__firecrawl_search,WebFetch category: AI & Machine Learning tags:

  • drone
  • slam
  • navigation
  • sensor-fusion
  • path-planning pairs-with:
  • skill: drone-inspection-specialist reason: Domain-specific inspection tasks
  • skill: physics-rendering-expert reason: Physics simulation for drone systems

Drone CV Expert

Expert in robotics, drone systems, and computer vision for autonomous aerial platforms.

Decision Tree: When to Use This Skill

User mentions drones or UAVs?
├─ YES → Is it about inspection/detection of specific things (fire, roof damage, thermal)?
│        ├─ YES → Use drone-inspection-specialist
│        └─ NO → Is it about flight control, navigation, or general CV?
│                ├─ YES → Use THIS SKILL (drone-cv-expert)
│                └─ NO → Is it about GPU rendering/shaders?
│                        ├─ YES → Use metal-shader-expert
│                        └─ NO → Use THIS SKILL as default drone skill
└─ NO → Is it general object detection without drone context?
        ├─ YES → Use clip-aware-embeddings or other CV skill
        └─ NO → Probably not a drone question

Core Competencies

Flight Control & Navigation

  • PID Tuning: Position, velocity, attitude control loops
  • SLAM: ORB-SLAM, LSD-SLAM, visual-inertial odometry (VIO)
  • Path Planning: A*, RRT, RRT*, Dijkstra, potential fields
  • Sensor Fusion: EKF, UKF, complementary filters
  • GPS-Denied Navigation: AprilTags, visual odometry, LiDAR SLAM

Computer Vision

  • Object Detection: YOLO (v5/v8/v10), EfficientDet, SSD
  • Tracking: ByteTrack, DeepSORT, SORT, optical flow
  • Edge Deployment: TensorRT, ONNX, OpenVINO optimization
  • 3D Vision: Stereo depth, point clouds, structure-from-motion

Hardware Integration

  • Flight Controllers: Pixhawk, Ardupilot, PX4, DJI
  • Protocols: MAVLink, DroneKit, MAVSDK
  • Edge Compute: Jetson (Nano/Xavier/Orin), Coral TPU
  • Sensors: IMU, GPS, barometer, LiDAR, depth cameras

Anti-Patterns to Avoid

1. "Simulation-Only Syndrome"

Wrong: Testing only in Gazebo/AirSim, then deploying directly to real drone. Right: Simulation → Bench test → Tethered flight → Controlled environment → Field.

2. "EKF Overkill"

Wrong: Using Extended Kalman Filter when complementary filter suffices. Right: Match filter complexity to requirements:

  • Complementary filter: Basic stabilization, attitude only
  • EKF: Multi-sensor fusion, GPS+IMU+baro
  • UKF: Highly nonlinear systems, aggressive maneuvers

3. "Max Resolution Assumption"

Wrong: Processing 4K frames at 30fps expecting real-time performance. Right: Resolution trade-offs by altitude/speed:

AltitudeSpeedResolutionFPSRationale
<30mSlow1920x108030Detail needed
30-100mMedium1280x72030Balance
>100mFast640x48060Speed priority

4. "Single-Thread Processing"

Wrong: Sequential detect → track → control in one loop. Right: Pipeline parallelism:

Thread 1: Camera capture (async)
Thread 2: Object detection (GPU)
Thread 3: Tracking + state estimation
Thread 4: Control commands

5. "GPS Trust"

Wrong: Assuming GPS is always accurate and available. Right: Multi-source position estimation:

  • GPS: 2-5m accuracy outdoor, unavailable indoor
  • Visual odometry: 0.1-1% drift, lighting dependent
  • AprilTags: cm-level accuracy where deployed
  • IMU: Short-term only, drift accumulates

6. "One Model Fits All"

Wrong: Using same YOLO model for all scenarios. Right: Model selection by constraint:

ConstraintModelNotes
Latency criticalYOLOv8n6ms inference
BalancedYOLOv8s15ms, better accuracy
Accuracy firstYOLOv8x50ms, highest mAP
Edge deviceYOLOv8n + TensorRT3ms on Jetson

Problem-Solving Framework

1. Constraint Analysis

  • Compute: What hardware? (Jetson Nano = ~5 TOPS, Xavier = 32 TOPS)
  • Power: Battery capacity? Flight time impact?
  • Latency: Control loop rate? Detection response time?
  • Weight: Payload capacity? Center of gravity?
  • Environment: Indoor/outdoor? GPS available? Lighting conditions?

2. Algorithm Selection Matrix

ProblemClassical ApproachDeep LearningWhen to Use Each
Feature trackingKLT optical flowFlowNetClassical: Real-time, limited compute. DL: Robust, more compute
Object detectionHOG+SVMYOLO/SSDClassical: Simple objects, no GPU. DL: Complex, GPU available
SLAMORB-SLAMDROID-SLAMClassical: Mature, debuggable. DL: Better in challenging scenes
Path planningA*, RRTRL-basedClassical: Known environments. DL: Complex, dynamic

3. Safety Checklist

  • Kill switch tested and accessible
  • Geofence configured
  • Return-to-home altitude set
  • Low battery action defined
  • Signal loss action defined
  • Propeller guards (if applicable)
  • Pre-flight sensor calibration
  • Weather conditions checked

Quick Reference Tables

MessagePurposeFrequency
HEARTBEATConnection alive1 Hz
ATTITUDERoll/pitch/yaw10-100 Hz
LOCAL_POSITION_NEDPosition10-50 Hz
GPS_RAW_INTRaw GPS1-10 Hz
SET_POSITION_TARGETCommandsAs needed

Kalman Filter Tuning

MatrixHigh ValuesLow Values
Q (process noise)Trust measurements moreTrust model more
R (measurement noise)Trust model moreTrust measurements more
P (initial covariance)Uncertain initial stateConfident initial state

Common Coordinate Frames

FrameOriginAxesUse
NEDTakeoff pointNorth-East-DownNavigation
ENUTakeoff pointEast-North-UpROS standard
BodyDrone CGForward-Right-DownControl
CameraLens centerRight-Down-ForwardVision

Reference Files

Detailed implementations in references/:

  • navigation-algorithms.md - SLAM, path planning, localization
  • sensor-fusion-ekf.md - Kalman filters, multi-sensor fusion
  • object-detection-tracking.md - YOLO, ByteTrack, optical flow

Simulation Tools

ToolStrengthsWeaknessesBest For
GazeboROS integration, physicsGraphics qualityROS development
AirSimPhotorealistic, CV-focusedWindows-centricVision algorithms
WebotsMulti-robot, accessibleLess drone-specificSwarm simulations
MATLAB/SimulinkControl designNot real-timeController tuning

Emerging Technologies (2024-2025)

  • Event cameras: 1μs temporal resolution, no motion blur
  • Neuromorphic computing: Loihi 2 for ultra-low-power inference
  • 4D Radar: Velocity + 3D position, works in all weather
  • Swarm autonomy: Decentralized coordination, emergent behavior
  • Foundation models: SAM, CLIP for zero-shot detection

Integration Points

  • drone-inspection-specialist: Domain-specific detection (fire, damage, thermal)
  • metal-shader-expert: GPU-accelerated vision processing, custom shaders
  • collage-layout-expert: Report generation, visual composition

Key Principle: In drone systems, reliability trumps performance. A 95% accurate system that never crashes is better than 99% accurate that fails unpredictably. Always have fallbacks.

スコア

総合スコア

60/100

リポジトリの品質指標に基づく評価

SKILL.md

SKILL.mdファイルが含まれている

+20
LICENSE

ライセンスが設定されている

+10
説明文

100文字以上の説明がある

0/10
人気

GitHub Stars 100以上

0/15
最近の活動

3ヶ月以内に更新

+5
フォーク

10回以上フォークされている

0/5
Issue管理

オープンIssueが50未満

+5
言語

プログラミング言語が設定されている

+5
タグ

1つ以上のタグが設定されている

0/5

レビュー

💬

レビュー機能は近日公開予定です