diff --git a/README.md b/README.md index 91115b7..89c7c0c 100644 --- a/README.md +++ b/README.md @@ -120,9 +120,6 @@ A simple pipeline visualization is as follows: ![Grounded SAM 2 Tracking Pipeline](./assets/grounded_sam2_tracking_pipeline_vis.png) -> [!NOTE] -> We noticed that the SAM 2 video predictor also supports mask prompts. Using the SAM 2 mask prediction results should be a more straightforward approach, and we will explore this implementation in future release. - ### Grounded-SAM-2 Video Object Tracking Demo (with Grounding DINO 1.5 & 1.6) We've also support video object tracking demo based on our stronger `Grounding DINO 1.5` model and `SAM 2`, you can try the following demo after applying the API keys for running `Grounding DINO 1.5`: @@ -146,6 +143,7 @@ VIDEO_PATH = "./assets/hippopotamus.mp4" TEXT_PROMPT = "hippopotamus." OUTPUT_VIDEO_PATH = "./hippopotamus_tracking_demo.mp4" API_TOKEN_FOR_GD1_5 = "Your API token" # api token for G-DINO 1.5 +PROMPT_TYPE_FOR_VIDEO = "mask" # using SAM 2 mask prediction as prompt for video predictor ``` After running our demo code, you can get the tracking results as follows: