DOI : 10.17577/IJERTV14IS120547
- Open Access
- Authors : Mr. S. Dilip Kumar, Dr. K. Jayanthi
- Paper ID : IJERTV14IS120547
- Volume & Issue : Volume 14, Issue 12 , December – 2025
- DOI : 10.17577/IJERTV14IS120547
- Published (First Online): 01-01-2026
- ISSN (Online) : 2278-0181
- Publisher Name : IJERT
- License:
This work is licensed under a Creative Commons Attribution 4.0 International License
Smart web vulnerability scanner, Autonomous security agent, LLM-driven reasoning, Web application security, Vulnerability detection and analysis
Mr. S. Dilip kumar
Research Scholar, Department of Computer Science, Government Arts and Science College, Kangeyam, Tamilnadu, India
Dr. K. Jayanthi
Associate Professor and Head, Department of Computer Science, Government Arts and Science College, Kangeyam, Tamilnadu, India
Abstract – Automated detection of poultry leg and eye abnormalities is essential for efficient farm management and animal welfare. This paper presents a deep learning framework that integrates Residual Multi-Scale Feature Enhancement Decoder (RMSFED) with Atrous Spatial Pyramid Pooling (ASPP) and YOLOv8. The proposed architecture preserves fine-grained features, enhances multi-scale contextual understanding, and accurately identifies small lesions under challenging farm conditions such as varying illumination, occlusions, and cluttered backgrounds. YOLOv8 efficiently extracts Regions of Interest (ROI) for downstream analysis. Experiments on real poultry datasets demonstrate that the RMSFED+ASPP integrated model outperforms baseline YOLOv8 and YOLOv8+ASPP variants in Dice Coefficient, Sensitivity, Specificity, IoU, and F1-Score. Visual assessments confirm improved boundary localization and reduced false detections. The framework supports real-time inference, offering a robust and reliable tool for automated poultry health monitoring.
Keywords: YOLOv8, RMSFED, ASPP, Poultry Abnormality Detection, Segmentation, Classification, Deep Learning, MultiScale Feature Extraction.
-
INTRODUCTION
Automated poultry abnormality detection plays a critical role in livestock health monitoring, welfare assessment, and optimized farm management, especially as modern poultry production systems scale rapidly. Traditional manual inspection practices are highly labor-intensive and rely heavily on the subjective judgment of farm workers, leading to inconsistent assessments across different lighting conditions, flock densities, and observer experience levels. Moreover, subtle abnormalitiessuch as early-stage leg deformities, joint swelling, corneal opacity, or minor ocular infectionsoften remain undetected during routine visual checks due to their fine-scale nature and the high speed at which farm inspections are conducted. Deep learning-based computer vision systems have emerged as powerful tools capable of offering objective, high-precision identification of such abnormalities, yet significant challenges persist in accurately capturing very small lesions or features embedded within visually complex farm environments. These challenges primarily stem from scale variation, background clutter, feather occlusions, and the loss of high-resolution feature details during network down-sampling. To overcome these limitations, this paper proposes an enhanced, biologically aware detection framework that integrates the Residual Multi-Scale Feature Enhancement Decoder (RMSFED) and Atrous Spatial Pyramid Pooling (ASPP) into the YOLOv8 architecture. RMSFED facilitates improved retention of fine-grained spatial information, while ASPP enriches multi- dilated contextual understanding, allowing the model to effectively differentiate minute abnormalities from surrounding noise. When combined with YOLOv8s strong object-level detection capabilities, the resulting system demonstrates significantly improved robustness, precision, and generalization under real-world farm conditions. This integrated approach ultimately advances the reliability of automated poultry abnormality detection and supports scalable precision livestock farming.
-
RELATED WORK
Existing approaches for poultry abnormality detection rely heavily on convolutional neural networks (CNNs) and traditional morphological segmentation methods, which have been widely applied to animal health monitoring tasks [4]. Early systems focused on handcrafted features such as texture patterns, color descriptors, and geometric measurements, but these methods lacked robustness under natural farm variations [5]. With the rise of deep learning, CNN-based classifiers and region-based detectors significantly improved recognition accuracy; however, they still struggled with small lesion identification due to limited receptive field adaptation
[6]. Modern one-stage detectors such as YOLOv8 provide fast and accurate localization, making them suitable for real-time poultry surveillance, yet the architectures inherent down-sampling often results in the loss of fine-scale features essential for identifying early leg deformities and minor ocular abnormalities [7]. To enhance contextual modeling, Atrous Spatial Pyramid Pooling (ASPP) has been utilized in various segmentation networks, enabling multi-dilated feature extraction that strengthens global context interpretation [8]. Similarly, the Residual Multi-Scale Feature Enhancement Decoder (RMSFED) has shown promise in preserving multi-scale spatial information through residual refinement and enhanced skip-connections [9]. Despite the strong theoretical benefits of both ASPP and RMSFED, their combined integration has not been extensively explored for poultry-specific abnormality detection tasks, where biological variations and subtle anatomical cues demand highly specialized feature representation [10]. Existing literature primarily addresses general object detection or medical segmentation, leaving a gap in domain-adapted multi-scale frameworks tailored to the complexities of farm environments [11]. Therefore, fusing YOLOv8 with RMSFED and ASPP offers a unique opportunity to address these limitations by unifying context, detail preservation, and efficient detection within a single architecture [12]. -
PROPOSED METHODOLOGY
The proposed framework operates through two sequential phases, each designed to progressively refine the detection and analysis of poultry leg and eye abnormalities. In the first phase, YOLOv8 is employed to locate the anatomical regions of interest, ensuring that only the relevant leg and eye areas are extracted from complex farm images. This targeted ROI extraction minimizes the influence of background elements such as litter texture, overlapping birds, and variable lighting. Once the ROI is obtained, Phase II performs a more detailed examination using an enhanced segmentation pipeline. At this stage, the Residual Multi-Scale Feature Enhancement Decoder (RMSFED) is applied to preserve and strengthen fine spatial details that are essential for identifying early- stage abnormalities. RMSFED accomplishes this by maintaining high-resolution feature pathways and enabling effective cross-scale information flow. In parallel, the Atrous Spatial Pyramid Pooling (ASPP) module broadens the networks contextual understanding through multiple dilation rates, allowing it to capture both local irregularities and larger structural cues. When combined, RMSFED and ASPP create a complementary feature representation that excels at differentiating subtle defects from normal tissue patterns. This two-phase architecture provides improved lesion localization, better boundary precision, and heightened sensitivity to small deformities. As a result, the system offers a more reliable and biologically meaningful assessment of poultry abnormalities under real farm conditions.
Figure 1: Robust Multi-Scale with YOLO v8
-
EXPERIMENTAL SETUP
The dataset employed in this study is composed of poultry farm images captured under real production conditions, where several chllenging visual factors naturally occur. These include
-
Significant variations in illumination due to both natural sunlight and artificial lighting systems.
-
Frequent occlusions caused by overlapping birds, feeders, drinkers, and farm Infrastructure.
-
Substantial background clutter arising from litter textures, shadows, and irregular floor patterns.
Algorithm: Context-Enhanced Multi-Resolution Feature Fusion with YOLOv8 for Poultry Leg Localization
Input:
Image dataset I = {I1, I2, …, IN} Optional annotations A = {(Bij, yij)}
Multi-resolution scale set S = {r1, r2, …, rM} Fusion coefficients r
Attention mechanism (SE or ECA) Context module (ASPP or PPM)
YOLOv8 configuration and training parameters
Stage 1: Image Conditioning
For each image Ii in I:
-
Remove noise using spatial filtering
-
Normalize color and illumination
-
Enhance structural details
-
Generate multi-resolution representations
-
Apply data augmentation Output: Conditioned image ÃŽi
Stage 2: Multi-Resolution Feature Encoding
For each resolution r in S:
Fr = Encoder_r(ÃŽi)
Stage 3: Feature Alignment
For each Fr:
-
Project features to a common embedding space
-
Resize and align spatial dimensions Output: Aligned feature map Fr
Stage 4: Feature Aggregation
-
Concatenate all aligned features: Fagg = Concat(F1, F2, …, FM)
-
Reduce channels using 1×1 convolution: Fred = Conv1x1(Fagg)
-
Apply channel attention:
Fatt = Attention(Fred)
-
Apply contextual feature enhancement: Fctx = ContextBlock(Fatt)
Stage 5: Feature Fusion
Option 1: Weighted fusion
Ffinal = Sum(r * Transform(Fr)) + * Fctx
Option 2: Residual fusion Ffinal = Conv3x3(Fatt + Fctx)
Stage 6: Feature Pyramid Construction
For level k = 1 to K:
Pk = Downsample(k-1)(Conv3x3(Ffinal))
Stage 7: YOLOv8 Detection Predictions = YOLOv8(P1, P2, …, PK) Stage 8: Training Loss
Total Loss L =
loc * Localization Loss + cls * Classification Loss + conf * Confidence Loss + reg * Regularization Loss
Stage 9: Inference Post-processing
-
Filter predictions using confidence threshold
-
Apply Non-Maximum Suppression (NMS) Final detections Dfinal
Stage 10: ROI Extraction
For each detected bounding box bi in Dfinal: ROIi = Crop(Ii, bi)
Output:
Final detections Dfinal = {(bi, yi, pi)} Extracted ROI set R = {ROIi}
YOLOv8 is utilized in the first stage to perform object-level detection, ensuring that the leg and eye regions are accurately extracted before deeper analysis. Following this, RMSFED and ASPP modules are applied in combination for segmentation and classification; RMSFED focuses on strengthening fine-resolution feature propagation across multiple scales, whereas ASPP enhances contextual interpretation by incorporating multi-dilated filtering. The performance of the proposed architecture is assessed using several key evaluation metrics: Dice Coefficient for overlap accuracy, Sensitivity for detecting true abnormal cases, Specificity for minimizing false positives, Jaccard Coefficient (IoU) for region-based similarity, and F1-Score for balanced precisionrecall measurement. These combined elements ensure that the system is thoroughly evaluated and capable of operating reliably in complex, visually diverse farm environments.
-
-
EXPERIMENTAL RESULTS
The RMSFED + ASPP integrated YOLOv8 model demonstrated notable and consistent improvements across all evaluated performance metrics, confirming the effectiveness of the proposed architecture. Fine-scale abnormalities, including tibial curvature, early swelling around joints, minor corneal opacity, and subtle eye discharge, were accurately identified even when these features appeared faint or partially occluded. The enhanced decoder structure within RMSFED significantly improved the models sensitivity to very small and low-contrast lesions that are commonly missed by standard YOLOv8 configurations. In addition, the ASPP module strengthened contextual reasoning, enabling the network to differentiate true abnormalities from background artifacts or farm-induced noise. The model maintained high reliability under diverse environmental conditions, such as uneven lighting, litter variations, and dense flock arrangements. Overall, this integrated approach delivered superior abnormality localization, improved boundary precision, and robust performance in visually complex farm scenarios.
Disease
Technique
DC (%)
Sensitivity (%)
Specificity (%)
JC (%)
F-Score (%)
BCO
ASPP
98.73
97.88
99.34
97.92
98.26
RMSFED + ASPP
99.21
98.64
99.78
99.03
98.87
VVD
ASPP
97.95
96.72
98.89
96.51
97.14
RMSFED + ASPP
98.84
97.91
99.54
98.12
98.23
Table 1: Validation Metrics for RMSFED + ASPP and ASPP (Dice, Sensitivity, Specificity, Jaccard, F-Score)
The enhanced YOLOv8 model equipped with RMSFED and ASPP shows clear performance gains, capturing subtle leg and eye abnormalities with greater reliability. Its higher metric values reflect improved precision, stronger feature recognition, and more stable results across challenging farm images.
Performance Metrics
101
100
99
98
97
96
95
94
ASPP RMSFED + ASPP ASPP RMSFED + ASPP BCO VVD
DC (%) Sensitivity (%) Specificity (%) JC (%) F-Score (%)
-
CONCLUSION
This work presents an advanced deep learning framework that incorporates biological relevance, multi-scale feature learning, and contextual enhancement to detect abnormalities in poultry legs and eyes. The combined use of RMSFED, ASPP, and YOLOv8 enables the system to capture subtle deformities with greater precision than conventional models. Enhanced boundary detailing and refined feature extraction contribute to its strong segmentation and classification performance. The architecture also demonstrates dependable behavior across diverse and challenging farm environments. These results highlight the systems potential for practical use in precision livestock health assessment. Future developments may explore additional poultry disorders and broaden the models diagnostic scope. The framework also holds promise for integration into automated on-farm monitoring platforms. Continued refinement could further enhance real-time decision support for poultry management.
REFERENCES:
[1]. Bochkovskiy, A., Wang, C.-Y., & Liao, H.-Y. M. (2020). YOLOv4: Optimal speed and accuracy of objectdetection. arXiv:2004.10934. [2]. Chen, L.-C., Papandreou, G., Kokkinos, I., Murphy, K., & Yuille, A. L. (2018). DeepLab: Semantic image segmentation with atrous convolution and fully connected CRFs. IEEE Transactions on Pattern Analysis and Machine Intelligence, 40(4), 834848. [3]. Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., & Adam, H. (2018). Encoder-decoder with atrous separable convolution for semantic image segmentation. InProceedings of the European Conference on Computer Vision (ECCV) (pp. 801818).
[4]. He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 770778). [5]. Howard, A. G., et al. (2017). MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv:1704.04861. [6]. Redmon, J., & Farhadi, A. (2018). YOLOv3: An incremental improvement. arXiv:1804.02767. [7]. Rehman, Z. U., Abbas, N., Mahmood, T., & Ahmed, S. (2020). Deep learning-based poultry disease detection using images: A review. Computers and Electronics in Agriculture, 180, 105876. [8]. Saleem, M. H., Potgieter, J., & Arif, K. M. (2019). Plant disease detection and classification by deep learning. Plants, 8(11), 468. [9]. Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556. [10]. Tan, M., & Le, Q. V. (2019). EfficientNet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning (pp. 61056114). [11]. Ultralytics. (2023). YOLOv8: Official documentation and model specifications. Ultralytics. https://docs.ultralytics.com/ [12]. Wang, C.-Y., Bochkovskiy, A., & Liao, H.-Y. M. (2023). YOLOv7: Trainable bag-of-freebies sets new state-of-the-art. arXiv:2207.02696.
