Orchestrating segment anything models to accelerate segmentation annotation on agricultural image datasets

Read the full article See related articles

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Increasingly many applications of machine vision and artificial intelligence (AI) can be observed in agriculture. Yet, high quality training data remains a bottleneck in the development of many AI solutions which is particularly true for image segmentation. Therefore, ARAMSAM (agricultural rapid annotation module based on segment anything models) was developed, a user interface orchestrating the pre-labelling capabilities of both the segment anything models (SAM 1, SAM 2) as well as conventional annotation tools. One in silico experiment on zero-shot performance of SAM 1 and SAM 2 on three unseen agricultural datasets and another experiment on hyperparameter optimization of the automatic mask generators (AMG) were conducted. In a user experiment, 14 agricultural experts applied ARAMSAM to quantify the reduction of annotation times. SAM 2 benefitted greatly from hyperparameter optimization of its AMG . Based on ground truth masks matched with predicted masks, the F 2 -score of SAM 2 improved from 0.05 to 0.74 while SAM 1 was improved from 0.87 to 0.93. The user interaction time could be reduced to 2.1 s/mask on single images (SAM 1) and to 1.6 s/mask on image sequences (SAM 2) compared to polygon drawing (9.7 s/mask). This study demonstrates the potential of segment anything models as incorporated in ARAMSAM to significantly accelerate the process of segmentation mask annotation in agriculture and other fields. ARAMSAM will be released as open-source software (AGPL-3.0 license) at https://github.com/DerOehmer/ARAMSAM.

Article activity feed