Temporal Synchronization of Multi-View Video for Cattle Movement Analysis in Dynamic Farm Settings
Confirmed Presenter: Houda Orchi, Computer Science Department, Université du Québec à Montréal, Montreal, QC H3C 3P8, Canada, Canada
Room: 520c
Format: In Person
Authors List: Show
- Houda Orchi, Computer Science Department, Université du Québec à Montréal, Montreal, QC H3C 3P8, Canada, Canada
- Abdoulaye Baniré Diallo, Computer Science Department, Université du Québec à Montréal, Montreal, QC H3C 3P8, Canada, Canada
- Elsa Vasseur, Department of Animal Science, McGill University, Sainte-Anne-de-Bellevue, QC H9X 3V9, Canada, Canada
- Halima Elbiaze, Computer Science Department, Université du Québec à Montréal, Montreal, QC H3C 3P8, Canada, Canada
- Essaid Sabir, Computer Science Department, Université du Québec à Montréal, Montreal, QC H3C 3P8, Canada, Canada
- Mohammed Sadik, Department of Electrical Engineering, NEST Research Group ENSEM,Hassan II University, Casablanca, Morocco, Canada
Presentation Overview: Show
Synchronizing and aligning multi-angle video footage is an intricate task in computer vision, especially in complex environments like barn settings, where observing key behavioral events is critical for understanding dairy cow behavior. Various methods exist to address this challenge, but each has limitations. Frame-by-frame analysis, for instance, is too slow and error-prone for real-time use. Automated techniques like SIFT and SURF are ineffective in low-light or cluttered settings. Additionally, advanced SSM methods falter with dynamic textures and complex movements. To tackle these issues, we introduce a novel synchronization framework consisting of three main components.
First, it leverages YOLOv8-Oriented Bounding Boxes to extract bounding boxes that identify cow movements across videos, forming the cornerstone to construct self-similarity matrices that enable sophisticated comparisons of video sequences, by analyzing similarities through the dynamic positions and trajectories of the identified cows. Second, we enrich these matrices by incorporating advanced descriptors like the Histogram of Oriented Gradients (HOG) and optical flow. HOG aids in differentiating cow behaviors by analyzing edge directions and strengths while optical flow tracks pixel motion across frames. Third, we employ Dynamic Time Warping to identify minimal distances.
Our experiments involve real-time videos captured on the farm, comprising 50 recordings totaling nearly 1500 hours, meticulously curated and annotated by animal scientists. The synchronization quality is examined using assessment metrics like Earth Mover's Distance, Matched Frame Rate, and Mean Temporal Error. This evaluation confirms our proposed framework's effectiveness, demonstrating its potential to revolutionize livestock management and enhance animal welfare through synchronized multi-camera analysis.