A Simple Baseline for Zero-shot Semantic Segmentation with ...
arxiv.org › abs › 2112Dec 29, 2021 · In this paper, we target for zero-shot semantic segmentation, by building it on an off-the-shelf pre-trained vision-language model, i.e., CLIP. It is difficult because semantic segmentation and the CLIP model perform on different visual granularity, that semantic segmentation processes on pixels while CLIP performs on images.
TAO Pretrained Semantic Segmentation | NVIDIA NGC
catalog.ngc.nvidia.com › orgs › nvidiangc registry model list nvidia/tao/pretrained_semantic_segmentation:* Download the model: ngc registry model download-version nvidia/tao/pretrained_semantic_segmentation: --dest Instructions to run the sample notebook. Get the NGC API key from the SETUP tab on the left. Please store this key for future use. Detailed instructions can be found here