SAM is the revolutionary image segmentation model designed to improve annotation speed and quality in the world of computer vision.
In computer vision, segmentation plays a critical role and is based on pixel classification and defining pixels that belong to specific objects within an image. This technique has numerous applications, from analyzing scientific imagery to editing photos.
Nonetheless, achieving accurate annotation through segmentation can be quite challenging. And building a segmentation model demands expertise, AI training infrastructure, and vast amounts of annotated data.
Facebook's SAM project tackles all these challenges head-on.
The aim behind SAM's design was to boost image segmentation speed and precision by introducing a new comprehensive model, trained on a record-breaking 1-Billion mask dataset -- the largest segmentation dataset ever.
And the goal was accomplished. With SAM, there's no need for specialized knowledge, high computing power, or custom data. The SAM model has it all covered. It performs object detection and generates masks for them in any image or video frame, even those it hasn't encountered before.
SAM can be utilized for various applications without additional training, showcasing its impressive zero-shot transfer capability. It can be employed for data annotation across various fields, from medical to retail to autonomous vehicles. We eagerly anticipate discovering all the potential uses and applications that have yet to be imagined.
Now let's see how to use SAM in CVAT. This integration is currently available in a self-hosted solution and coming soon to CVAT.ai cloud!
Note, that SAM is an interactor model, It means you can annotate by using positive and negative points.
The process is easy and described in the following video:
Or if you prefer text to the video, follow this instruction:
1. If necessary, follow the basic instructions to install CVAT with serverless functions support.
2. The model is available on both CPU and GPU. The second option is significantly quicker, but if you want to install a GPU version, please additionally set up NVIDIA container Toolkit.
3. To deploy the Segment Anything interactor just perform one of the following commands from the root CVAT directory on your host machine:
Open CVAT, create a task, open an annotation job, and go to AI Tools > Interactors. You will find the model in the drop-down list.
Begin the annotation process by selecting the foreground using left mouse clicks and removing the background with right mouse clicks. Once the annotation is complete, save the job, and you'll be able to export the annotated objects in various supported formats.
We are currently working on adding the Segment Anything Model to CVAT.ai cloud!
Stay tuned and follow the news here: