M2cai16-tool-locations
path: ./m2cai16-tool-locations train: images/train val: images/val nc: 16 names: ['grasper','scissors','hook','clipper','irrigator','specimen_bag','bipolar','hook_electrode','trocars','stapler','suction','clip_applier','vessel_sealer','ligasure','ultrasonic','other'] This guide gives you a production‑ready starting point for loading, visualizing, converting, and training on the dataset. Adjust class names and annotation JSON structure based on your exact dataset version.
yolo detect train data=m2cai16.yaml model=yolov8n.pt epochs=100 imgsz=640 Example m2cai16.yaml : m2cai16-tool-locations
This dataset is designed for (bounding boxes) in laparoscopic cholecystectomy videos. It contains annotations for 16 tools, including their positions in video frames. 1. Dataset Overview & Utility Purpose : Train object detection models (e.g., YOLO, Faster R-CNN, DETR) to locate surgical instruments in real-time. It contains annotations for 16 tools, including their
def __len__(self): return len(self.samples) def __len__(self): return len(self
def __init__(self, root_dir, transform=None): self.root_dir = root_dir self.transform = transform self.samples = [] # Collect all (frame_path, annotation_path) pairs ann_dir = os.path.join(root_dir, 'annotations') for ann_file in os.listdir(ann_dir): if not ann_file.endswith('.json'): continue ann_path = os.path.join(ann_dir, ann_file) video_id = ann_file.replace('.json', '') frame_dir = os.path.join(root_dir, 'frames', video_id) with open(ann_path, 'r') as f: annotations = json.load(f) for frame_name, boxes_info in annotations.items(): frame_path = os.path.join(frame_dir, frame_name) if os.path.exists(frame_path): self.samples.append((frame_path, boxes_info))
boxes = target['boxes'].int() labels = target['labels'] class_names = dataset.CLASSES