Pigeon 3D Tracking
3D-POP Dataset
Using a large motion tracking system and pigeons with attached head and backpack markers, we tracked the fine-scale position and orientation of pigeon individuals. By finding the relative position of markers to keypoints of interest (e.g eyes, beak) by annotating 5-10 frames, we propagated keypoints using motion tracking information, allowing us to produce annotations for bounding box, individual trajectories, 2D and 3D keypoints for groups of 1, 2,5 and 10 pigeons.
3D-Muppet: 3D Multi-Pigeon Pose Estimation and Tracking
Using the 3D-POP dataset, we trained 2D keypoint detectors and triangulated 2D postures into 3D. The approach worked well and
yielded high accuracy (median 6.9mm keypoint error). We also show that training
a model with single pigeon data also works with multiple pigeons, for future
biologists who can potentially avoid annotating multi-animal data, which is
much more labour intensive than labelling single individuals.
Pigeons Everywhere: 2D postures of pigeons in all environments
Finally, we show that a model trained using data collected in captivity (3D-POP) can also work in pigeons in the wild, allowing 2D and 3D posture estimation of multiple individuals in the field. Using the segment anything model, we trained a 2D keypoint detection model of masked pigeons to remove the influence of the background. During inference, we first used a pre-trained MaskRCNN model to get pigeon masks, and managed to get 2D keypoint detection without additional annotations for training.