Rumination plays a pivotal role in assessing the health status of ruminants. However, conventional contact devices such as ear tags and pressure sensors raise animal welfare concerns during rumination behavior detection. Deep learning offers a promising solution for non-contact rumination recognition by training neural networks on datasets. We introduce UD-YOLOv5s, an approach for bovine rumination recognition that incorporates jaw skeleton feature extraction techniques. Initially, a skeleton feature extraction method is proposed for the upper and lower jaws, employing skeleton heatmap descriptors and the Kalman filter algorithm. Subsequently, the UD-YOLOv5s method is developed for rumination recognition. To optimize the UD-YOLOv5s model, the traditional intersection over the union loss function is replaced with the generalized one. A self-built bovine rumination dataset is used to compare the performance of three deep learning techniques: mean shift algorithm, mask region-based convolutional neural network, and you only look once version 3 (YOLOv3). The results of the ablation experiment demonstrate that UD-YOLOv5s achieves impressive precision (98.25%), recall (97.75%), and a mean average precision of 93.43%. We conducted a generalization performance evaluation in a controlled experimental environment to ensure fairness, indicating that UD-YOLOv5s converges faster than other models while maintaining comparable recognition accuracy. Moreover, our work reveals that when convergence speed is equal, UD-YOLOv5s outperforms other models regarding recognition accuracy. These findings provide robust support for accurately identifying cattle rumination behavior, showcasing the potential of the UD-YOLOv5s method in advancing ruminant health assessment. |
1.IntroductionRuminating is a distinctive digestive process observed in cattle, characterized by repetitive chewing, swallowing, and regurgitation, typically in peaceful surroundings such as grazing fields.1 The capability to recognize and track ruminating behavior in cattle presents an opportunity for enhancing livestock health management within the industry. Swiftly detecting and addressing any anomalies in cattle behavior can help mitigate potential health problems. Currently, diverse methodologies are employed to identify ruminating behavior in cattle, encompassing sensor technology, computer vision, behavior classification, and deep learning techniques.2 Using sensor technology is one of the most commonly employed methods for identifying rumination behavior in cattle. This approach involves detecting biological signals produced during rumination, such as changes in electrical potential resulting from chewing movements and sounds generated by gastrointestinal motility.3 While sensors offer reliable data, they require physical contact with animals and may raise concerns regarding animal welfare. Another frequently used technique is computer vision, which entails recording visual information using cameras and analyzing and processing video frames to extract rumination features. For example, contour extraction technology can be utilized to obtain cattle outlines.4 However, it is essential to consider the changing rumination scenes and lighting conditions when utilizing this method. Behavior classification offers a promising means of identifying rumination behavior using machine learning techniques. This approach involves training a classifier to recognize rumination behavior automatically. Supervised learning algorithms such as -nearest neighbors, support vector machines, random forests, etc., can be employed to train the data, and the classifier’s performance can then be evaluated using test data. However, ensuring accurate results is crucial, as this method requires a precise dataset and significant manual labeling work. Deep learning skeleton extraction is an innovative technique that leverages deep learning methods to extract object skeletons from images. This process involves two main steps: object contour detection and skeleton extraction. Initially, deep learning algorithms are employed to detect the object’s contour line, which is the foundation for extracting its skeleton, considering its shape and spatial relationships within the image. Subsequently, data about position, posture, and state can be determined by analyzing the skeleton in the image. This approach significantly enhances the accuracy of machine vision systems. The most commonly utilized techniques for deep learning skeleton extraction include convolutional neural networks (CNN), deep neural networks (DNNs), and recurrent neural networks.5 During training, input pixel information undergoes multiple layers of convolution and pooling operations, creating high-level feature representations. For a given image, the network utilizes feedback to extract more stable feature representations, allowing for the construction of different types of deep learning models tailored to the specific requirements of the target task. You only look once version 5 (YOLOv5) is an advanced deep learning model designed for precise object detection, efficiently identifying object locations and classes in images with remarkable accuracy.6 YOLOv5 comprises four basic models: YOLOv5s, YOLOv5m, YOLOv5l, and YOLOv5x. Among these, YOLOv5s stands out with its compact architecture and narrow feature width, making it particularly suitable for applications requiring high detection accuracy. As an evolution of single-stage object detection models, YOLOv5s significantly enhances network performance, achieving superior algorithmic accuracy and speed. This improvement is achieved by introducing additional layers and convolutional kernels to improve feature extraction and understanding while utilizing the cross-entropy loss function to refine predictions for different targets. YOLOv5s has gained widespread adoption across various domains, including intelligent surveillance, smart agriculture, and automated monitoring, due to its exceptional accuracy, rapid processing capabilities, and real-time detection and tracking of multiple objects.7 This paper introduces upper and lower you only look once version 5 small (UD-YOLOv5s), an innovative method for recognizing cattle rumination behavior based on extracting upper and lower jaw skeleton features using the YOLOv5s network. This work is the first to combine the YOLOv5s network with upper and lower jaw skeleton features for cattle rumination behavior recognition. The key contributions of this research are as follows:
The rest of this paper is organized as follows: Sec. 2 presents an overview of the current state of research in this field. Section 3 details the basic architecture of UD-YOLOv5s and the related processing methods. Section 4 outlines the materials and techniques. An analysis of the experimental results is provided in Sec. 5. Finally, in Sec. 6, we draw conclusions based on the findings from this research. 2.Research Status2.1.Progress in Ruminant Identification TechnologyBorchers et al.8 conducted a comprehensive work on rumination behavior in cattle, employing six different tri-axial accelerometer technologies. These included the cattle manager sensor, used to monitor rumination and feeding time, and the intelligent bow sensor, used to track cattle and monitor their rumination behavior and time spent in feeding areas. Visual observation was utilized to generate behavior time lengths, recording the start and end times of rumination behaviors occurring within a day. In another work, Bishop-Hurley et al.9 employed a collar system equipped with tri-axial accelerometers and magnetometers to investigate feed additives’ impact on cattle feeding behavior. In addition, they analyzed and developed behavior models using multivariate time series data to address potential issues with the sensors. Arablouei et al.10 developed a specialized pipeline for pre-processing, feature extraction, and cattle behavior classification using measurement data. This pipeline was designed to address resource constraints. They collected data from 10 cattle using tri-axial accelerometer sensors on collar tags to record their rumination and other behaviors. Watt et al.11 utilized acoustic methods to measure rumination activity in cows and discovered a positive correlation between dry matter intake and rumination time. Furthermore, they conducted a 10-day baseline detection for frequency analysis of rumination time. Rombach et al.12 validated the effectiveness of the ruminant welfare system (RWS) in measuring cows’ rumination behavior during barn grazing and supplementation periods. They improved the algorithm used in the RWS system. They found that the two converters used in the experiment were better at distinguishing rumination and feeding behavior in cows during other activities. Braun et al.13 established feature pressure curves by comparing data obtained from pressure sensors with 24-h direct observation data, demonstrating the regularity of chewing during rumination and generating a uniformly regular waveform. The consistency of the results was confirmed by comparing the data obtained through direct observation and pressure sensors. Gregorini et al.14 conducted a 21-day strip grazing experiment on a perennial ryegrass pasture with 8 cows equipped with high resolution (HR) tags and rumination collars to record rumination behavior and chewing activity during rumination. They calculated and registered rumination time based on the mean interval time of chewing actions, ultimately verifying the cows’ daily rumination pattern under the constraints of the pasture’s time. Handcock et al.15 demonstrated the potential of a wireless sensor network to monitor cattle behavior by combining global positioning system collars with satellite imagery, providing high temporal resolution. Schirmann et al.16 evaluated changes in rumination and feeding behavior before and after calving using time standards. Zehner et al.17 developed a novel scientific detection device to automatically measure stable-fed cow rumination and feeding behavior using a universal algorithm based on animal-specific learning data, with two software versions for the system. Pereira et al.18 evaluated the correlation and differences between direct visual observation and sensor data analysis by recording data for 6 h using a trained observer and ear-tag accelerometer sensors. Ruuska et al.19 proposed a pressure sensor system to measure rumination time in cattle and validated its effectiveness through comparative experiments. Porto et al.20 proposed a computer vision-based automated detection system for free-stall resting behavior in cattle that utilizes a Viola–Jones algorithm for cow lying behavior detection. The system captures image data of the research area through a multi-camera recording system and validates the system’s effectiveness by comparing the detection results with visual recognition results. Yazdanbakhsh et al.21 developed an intelligent system for continuous monitoring of the health status of each animal in livestock using sensors installed on animals. The experiment showed that higher sensitivity and specificity could be achieved using an integrated classifier in the wavelet domain. Arcidiacono et al.22 proposed a model based on an acceleration thresholding algorithm to detect cattle behavior by predicting the threshold of behavior onset from data obtained under specific experiments in the barn where the cattle is housed, thus determining the behavior of the cattle. Viazzi et al.23 developed a movement model based on lame cow walking behavior, increasing both single threshold accuracy and actual positive rate by representing the behavioral features hierarchically, effectively improving the accuracy of rapid and precise identification of lame cow walking on the farm. Pahl et al.24 analyzed feeding characteristics recorded by weighing troughs and rumination time changes recorded by acoustic sensors based on the feeding characteristics before and after cow estrus. Lee et al.25 organized and analyzed the predicted physiological parameters and various data from a wearable wireless sensor system for cows. The sensors used in the system primarily focus on rumination behavior, including ear tags, collars, and reticulum boluses. Jonsson et al.26 introduced individual animal lying balance by combining information from step sensors and leg tilt sensors, deriving a new change detection scheme, and establishing a change detection algorithm to describe cow behavior state within a given time interval using a binary variable. Braun et al.27 employed an automated system to assess feeding and rumination variables in a cohort of 300 dairy cows over 24 h. The method utilized pressure sensors integrated into neck collars to capture jaw movements. Subsequently, it analyzed vital parameters such as the duration of rumination, number of chewing cycles, number of chews per cycle, and units of chews per cycle. These studies have contributed significantly to understanding cattle rumination behavior and have used various sensor technologies and data analysis techniques to achieve meaningful insights. 2.2.Advances in Deep Learning-Based Cattle Regurgitation Recognition TechnologyJiang et al.28 proposed implementing the fast layered YOLO version 3 (FLYOLOv3) network for detecting essential parts of cattle in complex scenes based on the filter layer. They integrated a customized filter layer with an average filtering algorithm and leaky rectified linear unit (leaky ReLU) function to reduce training interference. Shakeel et al.29 introduced an innovative behavior recognition and computing scheme to predict cattle behavior. The proposed method used a deep recurrent learning paradigm to cycle the recognition pattern and classify abnormal situations based on differentiated data patterns. Chen et al.30 evaluated the latest developments in computer vision methods for recognizing cattle behavior based on animal productivity, health, and welfare. They analyzed the effect of image segmentation, recognition, and behavior recognition using traditional computer vision and deep learning methods, listing the progress of crucial research in this field. Peng et al.31 combined long short-term memory (LSTM) networks to detect and identify cattle behavior using inertial measurement units and classify behaviors such as regurgitation and feeding. LSTM - recurrent neural network (LSTM-RNN) model training showed potential for development in cattle regurgitation detection. Tamura et al.32 investigated the correlation between cattle behavior and acceleration data collected using a tri-axis neck-mounted accelerometer. They proposed the feasibility of improving behavior classification accuracy through machine learning. By recording the characteristic acceleration waves of eating, rumination, and lying down during visual observation of cattle behavior, they combined the farm dataset for decision-tree learning and ultimately verified the accuracy of decision-tree knowledge. Fuentes et al.33 proposed a hierarchical cattle behavior recognition method based on deep learning, incorporating spatiotemporal information. The framework includes appearance features at the frame level and spatiotemporal information containing more contextual time features. The designed system detected and located rumination behavior in multiple cattle in video frame regions. The method was validated using datasets captured in both day and night environments, showing effective recognition of 15 types of hierarchical activities. McDonagh et al.34 continuously monitored 46 cattle and used image recognition technology to predict their behaviors. Non-local neural networks were trained and validated on video clips for each behavior, showing successful recognition and classification of behaviors over extended periods. Dutta et al.35 classified cattle behavior recorded by a collar system equipped with a tri-axial accelerometer and magnetometer using machine learning techniques. A hybrid framework was developed to work the natural structure of sensor data, and various classification methods were compared to verify the model’s superiority in representing rumination behavior. Chen et al.36 proposed an intelligent monitoring method for cattle rumination behavior based on video analysis. They used the MEAN-SHIFT algorithm to track the movement of the cattle’s lower jaw and extract the centroid trajectory curve of the cattle’s mouth movement from the video, enabling monitoring of cattle rumination behavior. Li et al.37 proposed a multi-target monitoring method based on optical flow and frame difference for cattle rumination behavior. Using optical flow and frame difference, they identified candidate rumination cattle oral regions and tracked the cattle’s mouth within the area, enabling automatic monitoring of the rumination cattle’s mouth area. Tamura et al.38 developed a method for detecting chewing speed using a tri-axial accelerometer connected to the cattle’s neck. Using the fast Fourier transform algorithm, they analyzed neck vibration as rumination movement and calculated chewing time. Dutta et al.39 proposed an intelligent IoT device for cattle monitoring to identify rumination behavior and collect rumination data through classification. The device was mounted on the cattle’s neck, and data were transmitted to IoT servers and a cellular global-local space modulation module. Xu et al.40 constructed an advanced instance segmentation framework MASK-RCNN to address the problem of cattle dataset occlusion and overlap in complex scenes. They compared classical algorithms and validated the best threshold and complete detection for cattle herd data in different situations, demonstrating the effectiveness of the proposed framework. These studies provide valuable insights into monitoring cattle behavior and have practical applications in the agricultural industry.41–43 2.3.SummaryThe present work introduces an innovative technique for bovine rumination identification that addresses various perspectives. While rumination detection techniques based on wearable devices have limitations and may not meet animal welfare requirements, methods using pressure sensors and tri-axial accelerometers to detect rumination tend for prioritizing data collection over accurate identification. In addition, some approaches rely on identifying bovine body or head regions, which may need more accuracy in rumination recognition. In contrast, our proposed method focuses on bovine rumination identification by extracting features from the upper and lower jawbone skeleton. Specifically, we utilize skeleton extraction technology to extract the upper and lower jawbone skeleton features from the bovine mouth region. These features are then processed using the UD-YOLOv5s network architecture, enhancing the identification process’s accuracy and efficiency. We optimize the loss function to improve performance. By adopting this method, we achieve more accurate and efficient bovine rumination identification compared to traditional techniques. It ensures higher precision and considers the animals’ welfare, making it a promising step toward effective and humane bovine rumination identification. 3.UD-YOLOV5s-Based Method for Identifying Cattle Regurgitation Behavior3.1.UD-YOLOv5s System ArchitectureThis paper introduces UD-YOLOv5s for accurately identifying cattle grazing behavior. The proposed approach leverages the YOLOv5s network to extract crucial skeletal features from upper and lower jawbones. We combine the SHD and KF algorithms to achieve this and process the original feature map. Subsequently, the focus block is utilized for image segmentation, and the output is unsampled and fed into the convolution layer. The feature fusion module combines these features, subsequently connected to the pre-processing layer. Finally, we optimize the output using the GIoU loss function. The overall architecture of our system is shown in Fig. 1. The ruminant behavior of cattle primarily manifests through the skeletal movements of their upper and lower jaw regions. In this paper, we extract the skeleton features of the upper and lower jaw regions separately to discriminate ruminant behavior accurately. Specifically, the upper jaw region exhibits linear movement during cow rumination. We employ the SHD algorithm to extract the upper jaw skeleton structure from the cow’s image. The SHD algorithm is proficient in extracting the primary structural lines of objects by refining the edges of the image, thereby obtaining the skeleton features of the upper jaw. For the image of the cow’s upper jaw, the SHD algorithm effectively simplifies the complex edge structure into skeleton lines, thereby extracting the main structural features of the upper jaw. In addition, we utilize the KF, a filtering algorithm for state estimation while extracting the cow’s jaw skeleton features. The KF algorithm estimates the state of the jaw based on observed jaw position and motion trajectory, subsequently extracting the jaw’s motion features and skeleton information. The upper and lower jaw skeleton feature extraction captures distinct aspects of the cow’s ruminant behavior. Precisely, the upper jaw skeleton feature reflects the morphological characteristics of the cow’s mouth while the lower jaw skeleton feature reveals the movement characteristics of the cow’s lower jaw. A more comprehensive and integrated representation of ruminant behavioral features is achieved by fusing these two features. As a target detection network, the YOLOv5s network can accept multi-channel inputs. Therefore, the upper and lower jaw skeleton features can be effectively utilized as network inputs to achieve a comprehensive feature representation and analysis. This enhances the accuracy and effectiveness of our method for accurately identifying cattle grazing behavior. 3.2.UD-YOLOv5s Method Design3.2.1.Upper and lower jaw skeleton feature extractionSkeleton extraction is a valuable technique to detect essential points within an image and connect them sequentially, forming skeleton information. In the specific case of the cattle mouth region, the skeleton can be divided into two parts: the upper jaw and the lower jaw skeletons. The area of interest is first segmented into the upper and lower jaw regions to label the feature points in the original image. The feature points in the upper jaw are denoted as U1 to U6, corresponding to their coordinate positions starting from the left of the mouth area. Similarly, the feature points in the lower jaw are labeled as D1 to D5. Each pair of feature points between the upper and lower jaw is represented by a unit vector, such as U1→D1, connecting the respective points. However, due to the absence of a corresponding feature point for U6 in the lower jaw region, it is excluded from the skeleton. Figure 2 visually shows the labeled vital points in the cattle’s mouth region, offering an overview of the identified feature points. In this paper, we have employed the SHD detector, a highly accurate tool with a 90.90% success rate in identifying crucial points on the head, showcasing exceptional detection precision. To tackle the challenges posed by demanding conditions such as occlusion caused by cattle enclosures, complex environments, and overlapping bodies, we have innovatively utilized an 8-layer stacked hourglass network (8SHD). This network effectively transforms the original image into a feature map, creating a robust skeletal extraction model. The architecture of this model is represented in Fig. 3. This paper presents an advanced algorithm designed to extract crucial features of the mandible skeleton in cattle, which is vital in complex motion patterns. The method involves a multi-step process to ensure accuracy and reliability. First, the algorithm filters out low-confidence vital points, ensuring that only the most reliable maxillary structure critical points are considered. This selection is based on whether the output heat map exceeds a predefined threshold. An elliptical curve is employed to approximate the mandibular region, which aids in accurately capturing its shape. The algorithm then proceeds to identify vital points based on specific characteristics of the mandibular skeleton. These points serve as the foundation for further analysis. Next, a predictive algorithm is utilized to set key point features, refining the representation of essential points in the mandible. The algorithm considers the dynamic nature of the mandibular motion sequence, treating it as a dynamic system. During the opening and closing of the mandible, the left mouth edge point is selected as the original feature point, acting as a reference for the entire motion sequence. The curve formation is thoughtfully arranged by incorporating the mandibular skeleton feature and setting the adjacent bone distance to 0.1 cm per unit distance. In addition, the neighboring nodes of each unit period are considered adjacent skeleton feature points. By employing this sophisticated approach, the algorithm accurately models the motion pattern of the mandible. When is recorded as the initial frame , at this time, the jaw joint point gets the anchor frame and sits marked as . Currently, the next frame joint point position can be predicted by the KF algorithm as where denotes the state transfer matrix, and the site-tag of the next frame is predicted to be . The covariance matrix is calculated according to and prediction variable densities as where is denoted as the covariance matrix of , is represented as the covariance matrix of , is the process noise, and is the time stamp. The position of the detected node after processing is determined with the predicted position, and the updated state is calculated by Kalman gain. The coordinates of the matching node are replaced with the predicted coordinates. The gain is calculated asThe replacement state after gain is calculated as The error covariance is shown as where denotes the Kalman gain of frame ; denotes the updated lower jaw position; denotes the covariance matrix. We denote the obtained initial key point of the upper jaw as p1, the next predicted node of the frame as p2, and the adjacent algorithm-predicted nodes as p3 and p4. The initial key point of the lower jaw as p5, the next predicted vital point of the frame by the algorithm as p6, and the adjacent nodes as p7 and p8. The skeleton critical point representation is shown in Fig. 4.Information setting based on the jaw key point representation. Key point parameters are shown in Table 1. Table 1Key point information based on algorithm settings.
3.2.2.Design of UD-YOLOv5s networkThe YOLOv5s object detection algorithm is renowned for its exceptional detection accuracy and lightweight architecture. In this research, we introduce a groundbreaking approach called the UD-YOLOv5s network, specifically designed for detecting the mouth area of cattle and recognizing their rumination behavior. To achieve this goal, our methodology begins by resizing the input image to . Subsequently, we employ a slice operation to divide and connect the image, generating a processed image for further analysis. We incorporate spatial pyramid pooling and the convolutional block attention module (CBAM) to enhance the receptive field and feature expression. These components capture essential information from various spatial scales, optimizing the network’s performance. The data are then passed to the convolutional layer, where a bottleneck mechanism is employed for efficient feature extraction, thereby reducing the computational burden. The subsequent step involves feeding the up-sampling connected convolutional layer into the feature fusion module. This integration is executed bottom-up, employing the ReLU activation function, up-sampling connection, and GIoU loss function for effective optimization. Finally, the output of the feature map includes target bounding boxes, class probabilities, and confidence scores. A decoder completes this process, facilitating the precise identification and localization of cattle mouth regions. The architecture of the UD-YOLOv5s network is visually shown in Fig. 5, providing an overview of our proposed model’s intricate yet powerful design. In the context of bull’s mouth image analysis, addressing potential node mismatches that may occur when capturing images using a camera is essential. To ensure the accuracy and reliability of the skeleton features, we employ the object key point similarity (OKS) as a metric for evaluation, as shown in Eq. (6). OKS is a crucial measure to assess the similarity between the detected critical points in the image and the ground truth positions of these key points. Using OKS as an evaluation method, we can quantitatively gauge the alignment between the skeleton features extracted from the image and the expected positions of these features. The precise definition and calculation of OKS are shown in Eq. (6), allowing us to effectively evaluate and validate the correctness of the identified skeleton features compared to the ground truth. By incorporating the OKS evaluation method, we can enhance the robustness of our bull’s mouth image analysis, ensuring that the skeleton features accurately represent the underlying anatomical structures even in scenarios where node mismatch may occur during image capture where denotes the similarity of the key points of the maxillary skeleton, , represents the visibility of the critical points, the pixels of the individual feature points of the skeleton, the Euclidean distance between the actual coordinates of the critical points and the predicted coordinates, and the normalization factor of the key points. After inputting the regurgitated test set into the model, the average precision (AP) is calculated by calculating the OKS ratio to the skeleton individual node threshold, and determining whether the predicted target detection node coordinates match the actual coordinate points, if , the current image detection error is determined and the next image is input; , the skeleton node anchor frame is generated by the adaptive anchor frame mechanism, and the results are output to the backbone layer.The backbone network layer and neck network layer are our model’s core components and have been optimized using a lightweight approach. To enhance detection accuracy, we have incorporated the CBAM into the backbone layer for the skeleton nodes. CBAM comprises two main modules: the channel attention (CHN) module and the spatial attention (SPA) module. By leveraging CHN, the model can focus on relevant features within each channel, effectively improving the overall representation. Subsequently, the feature map is passed through the SPA, enabling the model to extract crucial spatial information, further enhancing feature extraction. For a visual representation of the CBAM attention mechanism, refer to Fig. 6. In CHN module, the original feature map , which is the product of height , width , and number of channels , is pooled to obtain the channel map, and then processed by multilayer perceptron (MLP) to obtain the feature weights, combined with the ReLU activation function to obtain the channel weight coefficients . The product of and is used as the scaled channel feature map , and is calculated as where represents the ReLU activation function, MLP the multilayer perceptron, and Pool the adaptive pooling operation. In the SPA module, the new feature map is pooled to obtain the channel map for stitching, and the spatial weight coefficient is obtained after layer convolution and ReLU activation function, and the product of and the feature map is used as the output feature map . The formula of is as in Eq. (8) where means the -layer convolution and the channel feature map. The filtered feature map is input to the Neck layer, as shown in Fig. 7.The neck layer employs a bottom-up path aggregation feature pyramid to transfer feature information efficiently. This pyramid facilitates the integration of feature maps corresponding to key points’ information. The process involves convolving the feature map of each skeleton key point with a step size of 1, which effectively fuses the features. Up-sampling techniques are utilized to accomplish the feature fusion, and the convolution module consists of a series of residual structures resource units. This integration process enhances the overall representation and prepares the feature information for the subsequent detector. Ultimately, the prepared feature information is fed into the detector, where regurgitated behavior recognition occurs, as illustrated in Eq. (9) where represents the feature information, Upsample the up-sampling operation, the convolution module integrated by residual structures, and Prediction the detector prediction operation. The neck layer architecture and output are shown in Fig. 8.The output tensor is partitioned into multiple small tensors, each corresponding to a network cell and its surrounding prediction frame. Within each small tensor, computations are performed to determine the target confidence, category probability, and bounding box coordinates. The final prediction takes the form of a three-dimensional tensor, where the first dimension represents the identification of each grid cell, the second dimension means the identification of each predefined frame, and the third dimension contains information related to target confidence, category probability, and bounding box coordinates. The upper and lower jaw skeleton feature extraction method is employed to obtain multiple anchor frames in the mouth region. This technique can derive anchor frames based on the skeleton features of the upper and lower jaws. Specifically, the center point of the upper jaw anchor frame serves as the base point, and it is connected to the center point of the lower jaw anchor frame. The Euclidean distance between these two anchor points is then calculated as where , , , and denote the horizontal coordinates of the center point of the upper and lower jaw anchor frames and the vertical coordinates of the center point of the upper and lower jaw anchor frames, respectively. Take the minimum value of , at this time frame is recorded as a chewing determination, regurgitation is consistent with chewing periodicity, so the determination of the number of regurgitation as shown in Eq. (11) where denotes the minima on the mastication curve in frame , the maxima in the adjacent next frame, and the mastication curve length. The final recognition prediction output is completed in the prediction layer.3.3.Optimization of UD-YOLOv5sRecognizing ruminant behavior poses a significant challenge due to an inherent dataset imbalance, with a limited number of ruminant samples compared to an abundance of non-ruminant instances. Traditional YOLOv5s rely on the IoU loss function, which needs to be more accurate in capturing the overlap between the target and authentic frames. This paper tackles this issue and aims to enhance model learning by introducing the GIoU loss function, offering a more precise measurement of frame overlap. We note that we retained all data in our work to maintain dataset integrity. 3.3.1.Loss weightingGIoU contains both localization error and classification error, so we increase the corresponding loss weights, and the localization error weight () and classification error () weights are calculated explicitly as where denotes the category and the number of positive samples of the sample. We see that the fewer the category samples, the larger the corresponding loss weight.3.3.2.Positive and negative sample samplingIn the context of a relatively small sample size of regurgitation behavior, the imbalance between positive and negative samples can adversely impact model training, leading to suboptimal results. We propose a positive and negative sample improvement strategy based on UD-YOLOv5s to address this issue. This strategy comprises three essential steps: redefining positive and negative samples, implementing a problematic sample mining approach, and enhancing the sample balancing strategy using the gradient harmonized (GHM) single stage detector loss for the GIoU loss function.
3.3.3.Dynamic weight adjustmentWe adjust the weight of the loss function for that class according to the magnitude of the calculated loss value, and set each sample weight in UD-YOLOv5s as where denotes the prediction result of sample , and are hyperparameters to implement the imbalance problem of GIoU loss function. The procedure of the UD-YOLOv5s learning algorithm is shown in Table 2.Table 2UD-YOLOv5s algorithm.
4.Methods and Materials4.1.Data AcquisitionThe experiment was conducted at a ranch in Yuan Yang, Henan Province, China, involving 120 mature Simmental cattle. To ensure comprehensive coverage of cattle activities, a camera was placed at a 45 deg angle above the ranch. Ranch caretakers meticulously screened the video data, and instances of cattle ruminating within 3 h of feeding at 5 AM and 5 PM were selected. Cattle were observed to be resting quietly during other periods. Subsequently, a frame-by-frame analysis was performed to process the video frames into image datasets. These resulting data files were accurately labeled based on their contents, with a resolution of . Both training and testing utilized the captured cattle behaviors. In addition, different degrees of occlusion were incorporated into the data collection process to reflect real-life farm conditions. The definition of cattle rumination is shown in Table 3, ensuring transparency and replicability of the work.44 Table 3Definition of cattle ruminant labeling.
4.2.Data Pre-ProcessingIn this work, the images utilized were impacted by low quality, mainly due to overlapping shapes caused by obstructions from the farm cattle pen, cattle movement, and poor adaptability to varying lighting conditions. To ensure data integrity, these low-quality images were excluded from the dataset. To enhance the generalization ability of the recognition network, data augmentation techniques were employed. Rotation and the mosaic data augmentation technique effectively augment the dataset. In addition, classic data augmentation techniques, such as image flipping, scaling, cropping, translation, noise addition, and contrast modification, were applied to diversify the dataset further. For a visual representation of the augmented dataset, refer to Fig. 9. These augmentation approaches collectively improve the model’s ability to recognize cattle behaviors amidst challenging environmental conditions and variations in image quality. Mosaic enhancement is a widely adopted image processing technique in object detection. In the YOLOv5s network, the cut and mix data augmentation algorithm is employed, effectively expanding the dataset and enhancing its diversity.45,46 Similarly, in UD-YOLOv5s, the stitcher method is incorporated as a data augmentation approach to tackle the issue of low accuracy in object detection caused by poor image quality resulting from various environmental factors during image capture.47 Figure 10 visually shows the process diagram for the mosaic enhancement technique employed in UD-YOLOv5s. By leveraging the mosaic enhancement and stitcher methods, the model’s ability to detect and recognize objects, even in challenging conditions with compromised image quality, is substantially improved. These techniques are pivotal in optimizing the performance of UD-YOLOv5s for object detection tasks. The mosaic-enhanced image of the data in this paper is shown in Fig. 11. In this context, consider Figs. 11(a1)–11(c1) as the original images and Figs. 11(a2)–11(c2) as the processed images. For data annotation, we utilized the labeling tool, and the resulting dataset is named the “cattle dataset.” This dataset comprises two distinct categories: “ruminate” and “other,” it encompasses a total of 1500 processed datasets. The data annotation details can be observed in Fig. 12. The use of the labeling tool ensures accurate and consistent annotation. At the same time, the division of the dataset into “ruminate” and “other” categories facilitates targeted analysis of cattle behavior. This comprehensive dataset lays the foundation for robust training and evaluation of the model, enabling effective detection and recognition of cattle activities. 4.3.Experimental Environment and Parameter SettingsThe experiments were conducted on a LINUX operating system, utilizing a robust hardware setup consisting of 16 GB of RAM, an NVIDIA GEFORCE RTX3070 graphics card for GPU acceleration, and a Core i7 CPU, along with an Intel(R) Core (TM) i7-10750H CPU@2.60 GHz 2.59 GHz processor for network training configuration. The model was built using the PyTorch deep learning framework, known for its efficiency and flexibility. Careful attention was given to the hyperparameter settings to ensure optimal performance during the training phase, as outlined in Table 4. These hyperparameter configurations play a crucial role in shaping the model’s learning process, enabling it to capture intricate patterns and features in the cattle dataset effectively. The robust hardware setup and well-tuned hyperparameters contribute significantly to the model’s accuracy and efficiency, enabling comprehensive training and evaluation of the object detection model for cattle behavior recognition. Table 4Hyperparameter setting.
To thoroughly evaluate the efficacy of the YOLOv5s network in detecting the upper and lower jaw regions of cattle during rumination behavior, we conducted an extensive rumination recognition detection experiment using the cattle dataset introduced in this paper. To ensure a fair and comprehensive comparison, we carefully selected MEAN-SHIFT, MASK-RCNN, and YOLOv3 as the control group networks, considering their similarities, versatility, and advancements in the field. MEAN-SHIFT, a clustering algorithm, segmented images by analyzing pixel color distribution. We extracted mouth features from video sequences and performed clustering operations using MEAN-SHIFT to recognize cattle rumination behavior occurring in the mouth area. MASK-RCNN, an enhanced algorithm capable of simultaneous object detection and instance segmentation, played a vital role in the experiment. We labeled the cattle rumination dataset and utilized pre-trained models to detect and locate the mouth area of the cattle. The region information was then fed into the MASK-RCNN network to perform mask segmentation, generating a segmentation mask for the cattle’s mouth. Subsequently, we extracted the feature vector and performed classification to identify cattle rumination behavior. The YOLOv3 algorithm involved video processing software to extract images containing cattle rumination behavior and conducted object detection to obtain the position information of the cattle’s mouth. We pulled relevant feature information, such as motion trajectory and color, and employed statistical analysis to determine the presence of rumination behavior. To ensure a fair comparison, we maintained consistent hyperparameters across all networks and presented identical parameter configurations for the four networks, as shown in Table 5. This approach allowed us to draw comprehensive conclusions and thoroughly evaluate the performance of UD-YOLOv5s in cattle rumination behavior recognition tasks from multiple perspectives. By conducting a rigorous comparative analysis, we aimed to provide robust insights into the strengths and weaknesses of each algorithm, ultimately highlighting the effectiveness of UD-YOLOv5s in this domain. Table 5Neural network parameter setting.
Since MEAN-SHIFT is a density-based unsupervised clustering algorithm, MASK-RCNN, YOLOv3, and UD-YOLOv5s are based on CNN architecture, the parameters of the four reconstructed networks are shown in Table 6. Table 6Reconfigured network parameters.
CSPNet, cross stage partial network. 5.Experiment5.1.UD-YOLOv5s Ablation ExperimentWe divided the dataset into “rumination” and “other,” represented by labels 0 and 1, respectively. For model training and validation, we utilized a 9:1 partition ratio. The training phase comprised 100 cycles to optimize model performance. To thoroughly evaluate the effectiveness of the jaw skeleton feature extraction method, we conducted ablation experiments under identical experimental conditions. In this paper, using the cattle dataset, we compare and validate the performance of four approaches: MEAN-SHIFT, MASK-RCNN, YOLOv3, and UD-YOLOv5s. The critical evaluation metrics for the models encompass precision (Pre), recall (Rec), frames per second (FPS), mean AP (mAP), model memory usage (MMU), and score. Pre denotes the ratio of samples predicted as positive by the genuinely positive model. In the context of cattle rumination recognition, it measures the accuracy of images correctly classified as rumination or non-rumination. Rec signifies the proportion of positive samples the model accurately predicts, indicating the number of samples correctly classified as positive by the classifier divided by the actual number of positive samples. It provides insight into how effectively rumination images are identified. The calculation method for the recall is illustrated in Eqs. (16) and (17). In this comprehensive evaluation, these metrics serve as key performance indicators, shedding light on the models’ abilities to detect cattle rumination behavior accurately. The higher the precision, recall, and score, the more reliable and effective the model recognizes cattle rumination patterns. In addition, the FPS, mAP, and MMU metrics contribute valuable insights into each model’s computational efficiency and memory utilization. These evaluations are crucial in determining the most suitable approach for cattle rumination recognition tasks where true positive (TP) represents the number of correctly identified positive cases, false positive (FP) indicates the number of FP claims, and false negative (FN) represents the number of FN claims. After conducting a thorough experimental analysis, UD-YOLOv5s demonstrates outstanding performance, achieving a remarkable 98.25% accuracy in bovine regurgitation recognition. This accuracy is 2.08% higher than that of YOLOv3, 7.22% higher than MASK-RCNN, and an impressive 10.18% higher than the MEAN-SHIFT algorithm. Regarding recall, UD-YOLOv5s showcases remarkable results, reaching a 97.75% recall rate. This is 4.07% higher than YOLOv3, 0.63% higher than MASK-RCNN, and 11.1% higher than the MEAN-SHIFT algorithm. The significant accuracy and recall improvements achieved by UD-YOLOv5s demonstrate its effectiveness and potential for real-world cattle rumination behavior detection tasks. The score is the weighted average of accuracy and recall, which can combine the model’s accurate prediction and completeness prediction and tells us that the model can accurately predict regurgitated pictures while avoiding missed detection as much as possible. The mAP indicates the degree of match between all boxes predicted by the model and the authentic boxes under different confidence thresholds. It is used to judge the overall performance of all models on regurgitated behavior recognition performance. It is calculated as shown in Eqs. (18) and (19) where AP denotes the average precision and indicates the total number of categories. The experimental results showed that UD-YOLOv5s reached 97.59%, 2.37% higher than YOLOv3, 3.27% higher than MASK-RCNN, and 11.92% higher than MEAN-SHIFT. Since the MEAN-SHIFT algorithm needs to dichotomize the results in the bovine rumination task, the mAP metric does not apply to this algorithm, and we did a mAP evaluation of the other three algorithms, which showed that UD-YOLOv5s reached 93.43%, 6.19% higher than YOLOv3, and 13.54% higher than MASK-RCNN. FPS refers to the number of frames a computer can process per second when analyzing video data. It serves as a crucial metric to gauge the real-time performance of a model in cattle ruminant recognition tasks, providing insights into the model network’s low latency and overall recognition improvement. A higher FPS value indicates faster model processing, resulting in shorter response times and more efficient real-time performance.On the other hand, MMU quantifies the memory size occupied by a trained deep-learning model. In the context of cattle ruminant recognition tasks, MMU serves as a measure of the model’s complexity. A smaller MMU implies a more compact model footprint, making it feasible to deploy the model on smaller devices for inference. This can help avoid overfitting and enhance the model’s generalization performance as it becomes more adaptable to various hardware configurations. The calculation of MMU is typically carried out following the equations provided in Eqs. (20) and (21) where denotes the time required to process a batch of images, Bit indicates the number of bytes per parameter data type, and denotes the number of model parameters. The experiments show that UD-YOLOv5s has a higher FPS with the same training time and a smaller model memory footprint with the same parameters and weights loaded. The results of the experiment are shown in Table 7.Table 7The results of the experiment.
5.2.Evaluation of the Generalization Effect of UD-YOLOv5sWe divided our self-built dataset into two main parts: the training dataset and the test dataset. To assess the generalization effect of the UD-YOLOv5s model, we conducted a thorough evaluation by comparing the accuracy and loss rate between these two sets. The dataset was further categorized into two groups: the regurgitated dataset and the other datasets which were used in the experiment. For the crossover subjects, we ensured that the training and test sets were appropriately balanced, consisting of 16,523 and 7152 samples, respectively. Similarly, for the crossover view, we maintained a balanced distribution of 14,236 samples in the training set and 6482 samples in the test set. Figure 13 illustrates the insightful results derived from our evaluation. The model’s accuracy exhibits a consistent upward trend as the number of training samples increases, reflecting the positive impact of additional training data on enhancing the model’s performance. The loss rates associated with the training and test sets in both cross-subject and cross-view iterative training are shown in Fig. 14. The figure demonstrates a notable pattern: as the number of training samples increases, the model’s loss rate consistently decreases. This observation highlights the beneficial effect of utilizing more training data, as it improves the model’s ability to minimize errors and optimize its predictions. This reduction in loss rate reflects the model’s enhanced learning and generalization capabilities, making it more adept at handling different scenarios and achieving better overall performance. 5.3.Performance ComparisonTo showcase the outstanding performance of the UD-YOLOv5s model in rumination recognition, we conducted a comprehensive comparative analysis involving three state-of-the-art models: MEAN-SHIFT, MASK-RCNN, and YOLOv3, all of which have demonstrated exceptional capabilities in rumination recognition research. MEAN-SHIFT focuses on tracking the mouth area and extracting mouth features to detect rumination patterns. MASK-RCNN, on the other hand, harnesses the power of the residual neural network to extract mouth area features and trains the algorithm, for instance, segmentation, enabling precise rumination recognition. As for YOLOv3, it processes the image using the backbone network, effectively fuses the features, and produces results through the prediction layer for rumination recognition. Since rumination recognition, which relies on the extraction of mandibular and maxillary skeleton features, is a multi-classification task, we employed the GIoU loss function for model training. To assess the superior convergence efficiency of the UD-YOLOv5s model, we meticulously compared the accuracy and loss values of MEAN-SHIFT, MASK-RCNN, and YOLOv3 at varying training cycle durations, as visually shown in Fig. 15. The results were resoundingly in favor of UD-YOLOv5s, exhibiting a remarkably faster convergence speed while maintaining the same level of recognition accuracy as the other models. When compared at the same convergence speed, UD-YOLOv5s achieved significantly higher recognition accuracy. Consequently, the UD-YOLOv5s model demonstrates unparalleled performance in rumination recognition accuracy when contrasted with the other models subjected to scrutiny. This work is based on the YOLOv5s, and the results of the experimental comparison with YOLOv5s are as follows in Table 8. Table 8YOLOv5s comparison experiment results.
Based on the accuracy, recall, mAP, and metrics, we conducted statistical tests on a self-constructed ruminant dataset using the following approach: Data preparation: the regurgitated and non-regurgitated datasets used in the experiment were selected as samples for the statistical test. Model evaluation: we compared and evaluated the different networks’ accuracy, recall, mAP, and scores. Testing method: considering the nature of the cattle ruminant dataset and the fulfillment of statistical assumptions, we chose the -test as the statistical testing method. Hypothesis determination: we established the null hypothesis () and alternative hypothesis (). The null hypothesis states that there is no significant difference between the indicators of the two models. In contrast, the alternative view suggests a significant difference exists between the indicators of the two models. Statistical calculations and significance tests: we performed -tests to calculate and assess the significance of the indicator differences. Based on the chosen method and the calculated results, the obtained -values were used to determine whether the differences were statistically significant. The experimental results demonstrated that the calculated -value was less than the chosen significance level (), allowing us to reject the null hypothesis and conclude that the observed differences are statistically significant. We analyzed the impact of UD-YOLOv5s on the interpretation of actual farm cattle regurgitation videos, as shown in Fig. 16. In the context of recognizing cattle regurgitation behavior in various scenarios, it is evident that the recognition method effectively identifies the cattle mouth area during regurgitation, along with other relevant behaviors, as highlighted by the red box in the figure. UD-YOLOv5s exhibits higher detection accuracy in scenarios without interference from external factors or the presence of cattle pens, overlap, and other complexities within the monitoring area. However, there is room for improvement in accuracy when dealing with such challenging environmental conditions. 6.ConclusionThis paper presents a novel approach, UD-YOLOv5s, for cattle rumination behavior recognition by leveraging upper and lower jaw skeleton feature extraction in combination with deep learning techniques. A series of experiments were conducted to assess the effectiveness and generalization of the proposed method using a self-built cattle dataset in a consistent experimental environment. Initially, model training and ablation experiments were performed to compare and analyze the performance of the upper and lower jaw skeleton feature extraction method in recognizing rumination behavior. Subsequently, the accuracy and loss rates were evaluated under cross-subject and cross-view conditions to verify the generalization effect of the UD-YOLOv5s model. Moreover, a comprehensive comparative experiment was conducted with other rumination behavior recognition methods, namely MEAN-SHIFT, MASK-RCNN, YOLOv3, and YOLOv5s. The results demonstrate that UD-YOLOv5s outperforms these methods’ recognition accuracy and performance under the same experimental conditions. In addition, it exhibits better loss rates on the cattle dataset. However, it was observed that as the size of the rumination dataset increases, the model’s speed may slow down during recognition and detection due to the global detection approach implemented in the algorithm, resulting in higher computational requirements. Addressing this limitation, future research will focus on enhancing the multi-object detection processing capabilities of UD-YOLOv5s to meet more extensive practical needs. In conclusion, this paper introduces UD-YOLOv5s, a robust cattle rumination behavior recognition method that combines skeleton extraction and deep learning techniques. The experiments and evaluations conducted on the self-built cattle dataset validate the effectiveness and generalization of the proposed method. Nonetheless, to further enhance its adaptability to complex environmental conditions, additional research efforts will be directed toward addressing challenges caused by varying lighting conditions, thus catering to a broader range of practical applications. Data AvailabilityThe data used to support this work are available from the corresponding author upon request. AcknowledgmentsConceptualization, formal analysis, methodology, resources and validation were completed by Guohong Gao, and data curation, investigation, methodology, review writing and editing was conducted by Chengchao Wang, Jianping Wang completed the data curation, resources, software, supervision, Yingying Lv conducted the Investigation and methodology, Qian Li completed the data curation and validation, Xueyan Zhang conducted the Investigation, methodology and validation, Zhiyu Li completed the data curation and supervision, Guanglan Chen completed the data curation and methodology. All authors have read and agreed to the published version of the manuscript. This work was partly supported by the Key Scientific and Technological Project of Henan Province (Grant Nos. 232102111128, 222102320181, and 212102310087), in part by the Innovation and Entrepreneurship Training Program of National College Students in China (Grant No. 202110467001), in part by the Major Special Project of Xinxiang City (Grant No. 21ZD003), and in part by the Key Scientific Research Projects of Colleges and Universities in Henan Province (Grant Nos. 23B520003 and 21A520001). The authors approved the version of the paper to be published. They agreed to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. The authors declare that there is no conflict of interest with respect to the publication of this paper. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. ReferencesA. Awad,
“From classical methods to animal biometrics: a review on cattle identification and tracking,”
Comput. Electron. Agric., 123
(8), 423
–435 https://doi.org/10.1016/j.compag.2016.03.014 CEAGE6 0168-1699
(2016).
Google Scholar
W. Andrew et al.,
“Automatic individual Holstein Friesian cattle identification via selective local coat pattern matching in RGB-D imagery,”
in IEEE Int. Conf. Image Process.,
484
–488
(2016). https://doi.org/10.1109/ICIP.2016.7532404 Google Scholar
L. E. Wallace et al.,
“Readability of thirteen different radio frequency identification ear tags by three different multi-panel reader systems for use in beef cattle,”
Prof. Anim. Sci., 24
(3), 384
–391 https://doi.org/10.15232/S1080-7446(15)30888-3
(2008).
Google Scholar
W. Andrew et al.,
“Visual localization and individual identification of Holstein Friesian cattle via deep learning,”
in IEEE Int. Conf. Comput. Vis. Workshops,
2850
–2859
(2017). https://doi.org/10.1109/ICCVW.2017.336 Google Scholar
O. Guzhva et al.,
“Now you see me: convolutional neural network-based tracker for dairy cows,”
Front. Robot. AI, 5 107 https://doi.org/10.3389/frobt.2018.00107
(2018).
Google Scholar
J. Redmon et al.,
“You only look once: unified, real-time object detection,”
in Proc. IEEE Conf. Comput. Vis. and Pattern Recognit.,
779
–788
(2016). https://doi.org/10.1109/CVPR.2016.91 Google Scholar
T. Uchino et al.,
“Individual identification model and method for estimating social rank among herd of dairy cows using YOLOV5,”
in IEEE 20th Int. Conf. Cogn. Inf. & Cogn. Comput.,
235
–241
(2021). https://doi.org/10.1109/ICCICC53683.2021.9811319 Google Scholar
M. R. Borchers et al.,
“A validation of technologies monitoring dairy cow feeding, ruminating, and lying behaviors,”
J. Dairy Sci., 99
(12), 7458
–7466 https://doi.org/10.3168/jds.2015-10843 JDSCAE 0022-0302
(2016).
Google Scholar
G. Bishop-Hurley et al.,
“An investigation of cow feeding behavior using motion sensors,”
in IEEE Int. Instrum. and Meas. Technol. Conf. (12MTC) Proc.,
1285
–1290
(2014). https://doi.org/10.1109/I2MTC.2014.6860952 Google Scholar
R. Arablouei et al.,
“In-situ classification of cattle behavior using accelerometry data,”
Comput. Electron. Agric., 183
(134), 106045 https://doi.org/10.1016/j.compag.2021.106045 CEAGE6 0168-1699
(2021).
Google Scholar
L. J. Watt et al.,
“Differential rumination, intake, and enteric methane production of dairy cows in a pasture-based automatic milking system,”
J. Dairy Sci., 98
(59), 7248
–7263 https://doi.org/10.3168/jds.2015-9463 JDSCAE 0022-0302
(2015).
Google Scholar
M. Rombach et al.,
“Evaluation and validation of an automatic jaw movement recorder (Rumi Watch) for ingestive and rumination behaviors of dairy cows during grazing and supplementation,”
J. Dairy Sci., 101
(3), 2463
–2475 https://doi.org/10.3168/jds.2016-12305 JDSCAE 0022-0302
(2018).
Google Scholar
U. Braun et al.,
“Evaluation of eating and rumination behavior in cows using a noseband pressure sensor,”
BMC Vet. Res., 9
(1), 1
–8 https://doi.org/10.1186/1746-6148-9-164
(2013).
Google Scholar
P. Gregorini et al.,
“Rumination behavior of grazing dairy cows in response to restricted time at pasture,”
Livestock Sci., 146
(1), 95
–98 https://doi.org/10.1016/j.livsci.2012.02.020
(2012).
Google Scholar
R. Handcock et al.,
“Monitoring animal behavior and environmental interactions using wireless sensor networks, GPS collars and satellite remote sensing,”
Sensors, 9
(5), 3586
–3603 https://doi.org/10.3390/s90503586 SNSRES 0746-9462
(2009).
Google Scholar
K. Schirmann et al.,
“Short communication: rumination and feeding behavior before and after calving in dairy cows,”
J. Dairy Sci., 96
(11), 7088
–7092 https://doi.org/10.3168/jds.2013-7023 JDSCAE 0022-0302
(2013).
Google Scholar
N. Zehner et al.,
“System specification and validation of a noseband pressure sensor for measurement of ruminating and eating behavior in stable-fed cows,”
Comput. Electron. Agric., 136
(29), 31
–41 https://doi.org/10.1016/j.compag.2017.02.021 CEAGE6 0168-1699
(2017).
Google Scholar
G. M. Pereira et al.,
“Technical note: validation of an ear-tag accelerometer sensor to determine rumination, eating, and activity behaviors of grazing dairy cattle,”
J. Dairy Sci., 101
(3), 2492
–2495 https://doi.org/10.3168/jds.2016-12534 JDSCAE 0022-0302
(2018).
Google Scholar
S. Ruuska et al.,
“Validation of a pressure sensor-based system for measuring eating, rumination and drinking behaviour of dairy cattle,”
Appl. Animal Behav. Sci., 174
(3), 19
–23 https://doi.org/10.1016/j.applanim.2015.11.005 AABSEV
(2016).
Google Scholar
S. M. C. Porto et al.,
“A computer vision-based system for the automatic detection of lying behaviour of dairy cows in free-stall barns,”
Biosyst. Eng., 115
(2), 184
–194 https://doi.org/10.1016/j.biosystemseng.2013.03.002
(2013).
Google Scholar
O. Yazdanbakhsh et al.,
“An intelligent system for livestock disease surveillance,”
Inf. Sci., 378
(15), 26
–47 https://doi.org/10.1016/j.ins.2016.10.026
(2017).
Google Scholar
C. Arcidiacono et al.,
“Moving mean-based algorithm for dairy cow’s oestrus detection from uniaxial-accelerometer data acquired in a free-stall barn,”
Comput. Electron. Agric., 175 105498 https://doi.org/10.1016/j.compag.2020.105498 CEAGE6 0168-1699
(2020).
Google Scholar
S. Viazzi et al.,
“Analysis of individual classification of lameness using automatic measurement of back posture in dairy cattle,”
J. Dairy Sci., 96
(1), 257
–266 https://doi.org/10.3168/jds.2012-5806 JDSCAE 0022-0302
(2013).
Google Scholar
C. Pahl et al.,
“Feeding characteristics and rumination time of dairy cows around estrus,”
J. Dairy Sci., 98
(1), 148
–154 https://doi.org/10.3168/jds.2014-8025 JDSCAE 0022-0302
(2015).
Google Scholar
M. Lee et al.,
“Wearable wireless biosensor technology for monitoring cattle: a review,”
Animals, 11
(10), 2779 https://doi.org/10.3390/ani11102779 AIWHAJ
(2021).
Google Scholar
R. Jonsson et al.,
“Oestrus detection in dairy cows from activity and lying data using on-line individual models,”
Comput. Electron. Agric., 76
(1), 6
–15 https://doi.org/10.1016/j.compag.2010.12.014 CEAGE6 0168-1699
(2011).
Google Scholar
U. Braun et al.,
“Evaluation of eating and rumination behaviour in 300 cows of three different breeds using a noseband pressure sensor,”
BMC Vet. Res., 11
(3), 1
–6 https://doi.org/10.1186/s12917-015-0549-8
(2015).
Google Scholar
B. Jiang et al.,
“FLYOLOv3 deep learning for key parts of dairy cow body detection,”
Comput. Electron. Agric., 166
(521), 104982 https://doi.org/10.1016/j.compag.2019.104982 CEAGE6 0168-1699
(2019).
Google Scholar
P. M. Shakeel et al.,
“A deep learning-based cow behavior recognition scheme for improving cattle behavior modeling in smart farming,”
Comput. Electron. Agric., 19 10053 https://doi.org/10.1016/j.iot.2022.100539 CEAGE6 0168-1699
(2022).
Google Scholar
C. Chen et al.,
“Behavior recognition of pigs and cattle: journey from computer vision to deep learning,”
Comput. Electron. Agric., 187 106255 https://doi.org/10.1016/j.compag.2021.106255 CEAGE6 0168-1699
(2021).
Google Scholar
Y. Peng et al.,
“Dam behavior patterns in Japanese black beef cattle prior to calving: Automated detection using LSTM-RNN,”
Comput. Electron. Agric., 169 105178 https://doi.org/10.1016/j.compag.2019.105178 CEAGE6 0168-1699
(2020).
Google Scholar
T. Tamura et al.,
“Dairy cattle behavior classifications based on decision tree learning using 3-axis neck-mounted accelerometers,”
Animal Sci. J., 99
(72), 589
–596 https://doi.org/10.1111/asj.13184 ANPREJ
(2019).
Google Scholar
A. Fuentes et al.,
“Deep learning-based hierarchical cattle behavior recognition with spatiotemporal information,”
Comput. Electron. Agric., 177
(852), 105627 https://doi.org/10.1016/j.compag.2020.105627 CEAGE6 0168-1699
(2020).
Google Scholar
J. McDonagh et al.,
“Detecting dairy cow behavior using vision technology,”
Agriculture, 11
(7), 675 https://doi.org/10.3390/agriculture11070675
(2021).
Google Scholar
R. Dutta et al.,
“Dynamic cattle behavioral classification using supervised ensemble classifiers,”
Comput. Electron. Agric., 111
(63), 18
–28 https://doi.org/10.1016/j.compag.2014.12.002 CEAGE6 0168-1699
(2015).
Google Scholar
Y. Chen et al.,
“Intelligent monitoring method of cow ruminant behavior based on video analysis technology,”
Int. J. Agric. Biol. Eng., 10
(82), 194
–202 https://doi.org/10.1016/j.compag.2022.106772
(2017).
Google Scholar
T. Li et al.,
“Tracking multiple target cows ruminant mouth areas using optical flow and inter-frame difference methods,”
IEEE Access, 7
(198), 185520
–185531 https://doi.org/10.1109/ACCESS.2019.2961515
(2019).
Google Scholar
T. Tamura et al.,
“Short communication: detection of mastication speed during rumination in cattle using 3-axis, neck-mounted accelerometers and fast Fourier transfer algorithm,”
J. Dairy Sci., 103
(8), 7180
–7187 https://doi.org/10.3168/jds.2019-17611 JDSCAE 0022-0302
(2020).
Google Scholar
D. Dutta et al.,
“MOOnitor: an IoT based multi-sensory intelligent device for cattle activity monitoring,”
Comput. Electron. Agric., 333
(292), 113271 https://doi.org/10.1016/j.sna.2021.113271 CEAGE6 0168-1699
(2022).
Google Scholar
B. Xu et al.,
“Automated cattle counting using Mask R-CNN in quadcopter vision system,”
Comput. Electron. Agric., 171
(598), 100530 https://doi.org/10.1016/j.compag.2020.105300 CEAGE6 0168-1699
(2020).
Google Scholar
H. Karmouni et al.,
“Fast computation of 3D discrete invariant moments based on 3D cuboid for 3D image classification,”
Circuits Syst. Signal Process., 40 3782
–3812 https://doi.org/10.1007/s00034-020-01646-w
(2021).
Google Scholar
M. Yamni et al.,
“Accurate 2D and 3D images classification using translation and scale invariants of Meixner moments,”
Multimedia Tools Appl., 80
(17), 26683
–26712 https://doi.org/10.1007/s11042-020-10311-y
(2021).
Google Scholar
M. Yamni et al.,
“Fast and accurate computation of 3D Charlier moment invariants for 3D image classification,”
Circuits Syst. Signal Process., 40
(12), 6193
–6223 https://doi.org/10.1007/s00034-021-01763-0
(2021).
Google Scholar
O. El Ogri et al.,
“3D image recognition using new set of fractional-order Legendre moments and deep neural networks,”
Signal Process. Image Commun., 98 116410 https://doi.org/10.1016/j.image.2021.116410 SPICEF 0923-5965
(2021).
Google Scholar
M. A. Tahiri et al.,
“Optimal 3D object reconstruction and classification by separable moments via the Firefly algorithm,”
in Int. Conf. Intell. Syst. and Comput. Vis. (ISCV),
1
–8
(2022). https://doi.org/10.1109/ISCV54655.2022.9806106 Google Scholar
O. El Ogri et al.,
“A new fast algorithm to compute moment 3D invariants of generalized Laguerre modified by fractional-order for pattern recognition,”
Multidimens. Syst. Signal Process., 32 431
–464 https://doi.org/10.1007/s11045-020-00745-w MUSPE5 0923-6082
(2021).
Google Scholar
H. Karmouni et al.,
“Fast computation of 3D Meixner’s invariant moments using 3D image cuboid representation for 3D image classification,”
Multimedia Tools Appl., 79 29121
–29144 https://doi.org/10.1007/s11042-020-09351-1
(2020).
Google Scholar
|