Skip to content

Latest commit

 

History

History

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

README.md

Data Collection

This application combines on-device open-vocabulary detection with an interactive frontend to auto-collect “snaps” (images + metadata) under configurable conditions.
It runs YOLOE on the DepthAI backend, and exposes controls in the UI for:

  • Selecting labels (by text or image prompt)
  • Adjusting confidence threshold
  • Enabling snap conditions (timed, no detections, low confidence, lost-in-middle)

Note: RVC4 standalone mode only.

Architecture

High-level

In-depth

Features

  • Class control
    • Update classes by text or upload an image to create a visual prompt
  • Confidence filter
    • Drop detections below a chosen threshold
  • Snapping (auto-capture)
    • Timed (periodic)
    • No detections (when a frame has zero detections)
    • Low confidence (if any detection falls below threshold)
    • Lost-in-middle (object disappears inside central area; edge buffer configurable)
    • Cooldowns reset when snapping is (re)started

Usage

Running this example requires a Luxonis RVC4 device connected to your computer. Refer to the documentation to set up your device if you haven't already.


Standalone Mode (RVC4)

To run the example in this mode, first install the oakctl tool using the installation instructions here.

The app can then be run with:

oakctl connect <DEVICE_IP>
oakctl app run .

Once the app is built and running you can access the DepthAI Viewer locally by opening https://<OAK4_IP>:9000/ in your browser (the exact URL will be shown in the terminal output).

Remote access

  1. You can upload oakapp to Luxonis Hub via oakctl
  2. And then you can just remotely open App UI via App detail