# Copyright (c) OpenMMLab. All rights reserved. import os from argparse import ArgumentParser from pathlib import Path import mmcv from mmdet.apis import inference_detector, init_detector from mmengine.config import Config, ConfigDict from mmengine.logging import print_log from mmengine.utils import ProgressBar, path from mmyolo.registry import VISUALIZERS from mmyolo.utils import switch_to_deploy from mmyolo.utils.labelme_utils import LabelmeFormat from mmyolo.utils.misc import get_file_list, show_data_classes def parse_args(): parser = ArgumentParser() parser.add_argument( 'img', help='Image path, include image file, dir and URL.') parser.add_argument('config', help='Config file') parser.add_argument('checkpoint', help='Checkpoint file') parser.add_argument( '--out-dir', default='./output', help='Path to output file') parser.add_argument( '--device', default='cuda:0', help='Device used for inference') parser.add_argument( '--show', action='store_true', help='Show the detection results') parser.add_argument( '--deploy', action='store_true', help='Switch model to deployment mode') parser.add_argument( '--tta', action='store_true', help='Whether to use test time augmentation') parser.add_argument( '--score-thr', type=float, default=0.3, help='Bbox score threshold') parser.add_argument( '--class-name', nargs='+', type=str, help='Only Save those classes if set') parser.add_argument( '--to-labelme', action='store_true', help='Output labelme style label file') args = parser.parse_args() return args def main(): args = parse_args() if args.to_labelme and args.show: raise RuntimeError('`--to-labelme` or `--show` only ' 'can choose one at the same time.') config = args.config if isinstance(config, (str, Path)): config = Config.fromfile(config) elif not isinstance(config, Config): raise TypeError('config must be a filename or Config object, ' f'but got {type(config)}') if 'init_cfg' in config.model.backbone: config.model.backbone.init_cfg = None if args.tta: assert 'tta_model' in config, 'Cannot find ``tta_model`` in config.' \ " Can't use tta !" assert 'tta_pipeline' in config, 'Cannot find ``tta_pipeline`` ' \ "in config. Can't use tta !" config.model = ConfigDict(**config.tta_model, module=config.model) test_data_cfg = config.test_dataloader.dataset while 'dataset' in test_data_cfg: test_data_cfg = test_data_cfg['dataset'] # batch_shapes_cfg will force control the size of the output image, # it is not compatible with tta. if 'batch_shapes_cfg' in test_data_cfg: test_data_cfg.batch_shapes_cfg = None test_data_cfg.pipeline = config.tta_pipeline # TODO: TTA mode will error if cfg_options is not set. # This is an mmdet issue and needs to be fixed later. # build the model from a config file and a checkpoint file model = init_detector( config, args.checkpoint, device=args.device, cfg_options={}) if args.deploy: switch_to_deploy(model) if not args.show: path.mkdir_or_exist(args.out_dir) # init visualizer visualizer = VISUALIZERS.build(model.cfg.visualizer) visualizer.dataset_meta = model.dataset_meta # get file list files, source_type = get_file_list(args.img) # get model class name dataset_classes = model.dataset_meta.get('classes') # ready for labelme format if it is needed to_label_format = LabelmeFormat(classes=dataset_classes) # check class name if args.class_name is not None: for class_name in args.class_name: if class_name in dataset_classes: continue show_data_classes(dataset_classes) raise RuntimeError( 'Expected args.class_name to be one of the list, ' f'but got "{class_name}"') # start detector inference progress_bar = ProgressBar(len(files)) for file in files: result = inference_detector(model, file) img = mmcv.imread(file) img = mmcv.imconvert(img, 'bgr', 'rgb') if source_type['is_dir']: filename = os.path.relpath(file, args.img).replace('/', '_') else: filename = os.path.basename(file) out_file = None if args.show else os.path.join(args.out_dir, filename) progress_bar.update() # Get candidate predict info with score threshold pred_instances = result.pred_instances[ result.pred_instances.scores > args.score_thr] if args.to_labelme: # save result to labelme files out_file = out_file.replace( os.path.splitext(out_file)[-1], '.json') to_label_format(pred_instances, result.metainfo, out_file, args.class_name) continue visualizer.add_datasample( filename, img, data_sample=result, draw_gt=False, show=args.show, wait_time=0, out_file=out_file, pred_score_thr=args.score_thr) if not args.show and not args.to_labelme: print_log( f'\nResults have been saved at {os.path.abspath(args.out_dir)}') elif args.to_labelme: print_log('\nLabelme format label files ' f'had all been saved in {args.out_dir}') if __name__ == '__main__': main()