Robust individual pig tracking
บทความในวารสาร
ผู้เขียน/บรรณาธิการ
กลุ่มสาขาการวิจัยเชิงกลยุทธ์
รายละเอียดสำหรับงานพิมพ์
รายชื่อผู้แต่ง: Jaoukaew, Aggaluck; Suwansantisuk, Watcharapan; Kumhom, Pinit
ผู้เผยแพร่: Institute of Advanced Engineering and Science
ปีที่เผยแพร่ (ค.ศ.): 2024
ชื่อย่อของวารสาร: IJECE
Volume number: 14
Issue number: 1
หน้าแรก: 279
หน้าสุดท้าย: 293
จำนวนหน้า: 15
นอก: 2088-8708
eISSN: 2722-2578
ภาษา: English-Great Britain (EN-GB)
บทคัดย่อ
The locations of pigs in the group housing enable activity monitoring and improve animal welfare. Vision-based methods for tracking individual pigs are noninvasive but have low tracking accuracy owing to long-term pig occlusion. In this study, we developed a vision-based method that accurately tracked individual pigs in group housing. We prepared and labeled datasets taken from an actual pig farm, trained a faster region-based convolutional neural network to recognize pigs’ bodies and heads, and tracked individual pigs across video frames. To quantify the tracking performance, we compared the proposed method with the global optimization (GO) method with the cost function and the simple online and real-time tracking (SORT) method on four additional test datasets that we prepared, labeled, and made publicly available. The predictive model detects pigs’ bodies accurately, with F1-scores of 0.75 to 1.00, on the four test datasets. The proposed method achieves the largest multi-object tracking accuracy (MOTA) values at 0.75, 0.98, and 1.00 for three test datasets. In the remaining dataset, the proposed method has the second-highest MOTA of 0.73. The proposed tracking method is robust to long-term occlusion, outperforms the competitive baselines in most datasets, and has practical utility in helping to track individual pigs accurately.
คำสำคัญ
Multi-object tracking Performance evaluation Pig detection Pig localization Pig tracking