import sys import argparse import numpy as np from pathlib import Path from matplotlib import pyplot as plt from typing import Any, Dict, List import torch from segment_anything import SamPredictor, sam_model_registry from utils import load_img_to_array, save_array_to_img, dilate_mask, \ show_mask, show_points def predict_masks_with_sam( img: np.ndarray, point_coords: List[List[float]], point_labels: List[int], model_type: str, ckpt_p: str, device="cuda" ): point_coords = np.array(point_coords) point_labels = np.array(point_labels) sam = sam_model_registry[model_type](checkpoint=ckpt_p) sam.to(device=device) predictor = SamPredictor(sam) predictor.set_image(img) masks, scores, logits = predictor.predict( point_coords=point_coords, point_labels=point_labels, multimask_output=True, ) return masks, scores, logits def build_sam_model(model_type: str, ckpt_p: str, device="cuda"): sam = sam_model_registry[model_type](checkpoint=ckpt_p) sam.to(device=device) predictor = SamPredictor(sam) return predictor def setup_args(parser): parser.add_argument( "--input_img", type=str, required=True, help="Path to a single input img", ) parser.add_argument( "--point_coords", type=float, nargs='+', required=True, help="The coordinate of the point prompt, [coord_W coord_H].", ) parser.add_argument( "--point_labels", type=int, nargs='+', required=True, help="The labels of the point prompt, 1 or 0.", ) parser.add_argument( "--dilate_kernel_size", type=int, default=None, help="Dilate kernel size. Default: None", ) parser.add_argument( "--output_dir", type=str, required=True, help="Output path to the directory with results.", ) parser.add_argument( "--sam_model_type", type=str, default="vit_h", choices=['vit_h', 'vit_l', 'vit_b'], help="The type of sam model to load. Default: 'vit_h" ) parser.add_argument( "--sam_ckpt", type=str, required=True, help="The path to the SAM checkpoint to use for mask generation.", ) if __name__ == "__main__": """Example usage: python sam_segment.py \ --input_img FA_demo/FA1_dog.png \ --point_coords 750 500 \ --point_labels 1 \ --dilate_kernel_size 15 \ --output_dir ./results \ --sam_model_type "vit_h" \ --sam_ckpt sam_vit_h_4b8939.pth """ parser = argparse.ArgumentParser() setup_args(parser) args = parser.parse_args(sys.argv[1:]) device = "cuda" if torch.cuda.is_available() else "cpu" img = load_img_to_array(args.input_img) masks, _, _ = predict_masks_with_sam( img, [args.point_coords], args.point_labels, model_type=args.sam_model_type, ckpt_p=args.sam_ckpt, device=device, ) masks = masks.astype(np.uint8) * 255 # dilate mask to avoid unmasked edge effect if args.dilate_kernel_size is not None: masks = [dilate_mask(mask, args.dilate_kernel_size) for mask in masks] # visualize the segmentation results img_stem = Path(args.input_img).stem out_dir = Path(args.output_dir) / img_stem out_dir.mkdir(parents=True, exist_ok=True) for idx, mask in enumerate(masks): # path to the results mask_p = out_dir / f"mask_{idx}.png" img_points_p = out_dir / f"with_points.png" img_mask_p = out_dir / f"with_{Path(mask_p).name}" # save the mask save_array_to_img(mask, mask_p) # save the pointed and masked image dpi = plt.rcParams['figure.dpi'] height, width = img.shape[:2] plt.figure(figsize=(width/dpi/0.77, height/dpi/0.77)) plt.imshow(img) plt.axis('off') show_points(plt.gca(), [args.point_coords], args.point_labels, size=(width*0.04)**2) plt.savefig(img_points_p, bbox_inches='tight', pad_inches=0) show_mask(plt.gca(), mask, random_color=False) plt.savefig(img_mask_p, bbox_inches='tight', pad_inches=0) plt.close()