Spaces:
Sleeping
Sleeping
metadata
title: Demo 2025
emoji: π
colorFrom: green
colorTo: red
sdk: docker
pinned: false
license: mit
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
Mission-guided detections
- Call
POST /process_videowith fieldsvideo(file),prompt(mission text), and optionaldetector(owlv2orhf_yolov8). The response is an MP4 stream containing the annotated frames. - Call
POST /mission_summarywith the same fields to receive JSON containing the structured mission plan plus the natural-language summary. This second endpoint isolates the OpenAI call, keeping the video response clean. - Under the hood the mission text still feeds into the OpenAI (
gpt-4o-mini) reasoning step that ranks the YOLO/COCO classes. Place your API key inside.envas eitherOPENAI_API_KEY=...orOpenAI-API: ...; the server loads it automatically on startup. - The top scored classes drive OWLv2 or YOLOv8 to align detections with the mission, and the detection log is summarized via another OpenAI call when requested.