Trusted Engineering Publisher
Serving Researchers Since 2012

Comprehensive AI based Player Performance Analyzer and Wellness Monitoring System

DOI : https://doi.org/10.5281/zenodo.19205115
Download Full-Text PDF Cite this Publication

Text Only Version

 

Comprehensive AI based Player Performance Analyzer and Wellness Monitoring System

Gangalekshmi K B, Irin Mathew, Krishnendu B Menon, Neha Prasannan

CS Dept, Toc H Institute Of Science and Technology, Kochi

Bini Mol N C

Asst.Prof of CS Dept, Toc H Institute Of Science and Technology, Kochi

Abstract – This research is about football analytics with the help of AI by enabling data-driven perfor- mance optimization, injury prevention, and tal- ent identification. Access to advanced analyt- ics technologies remains limited to elite teams due to cost and other technical barriers. This will present the idea, methodologies and im- plementation of an AI-Driven Sports Analytics and Wellness Platform for football players, as a mobile-first, cloud-based digital ecosystem that incorporates performance analysis, AI coach- ing, and holistic wellness monitoring. The sys- tem uses computer vision and machine learn- ing techniques, including YOLOv8 for player detection, DeepSORT for multi-object tracking, and pose estimation for action recognition, to extract player metrics such as speed, stamina, and technical accuracy from video data. These insights are presented through a personalized AI generated Player Performance Card, with an AI coaching module developed using Ran- dom Forest and Decision Tree model delivers position recommendations and adaptive train- ing plans for any specific player for the bet- terment of their performance. Then a parallel wellness engine which analyzes sleep, hydra- tion, nutrition, and recovery patterns to predict injury risk and support long-term player health is also presented. Along with the facility for players to showcase their talent through the AI generated performance card, the system nor- malizes professional-grade football intelligence and provides an affordable, accessible pathway for athlete development across all competitive levels.

Index Terms: AI, Computer Vision, Foot- ball Analytics, Health Monitoring,Performance Metrics.

  1. Introduction

    In todays digital world, Artificial Intelligence (AI) is changing the way many industries work, and sports is no different. Football, in par- ticular, has started using technology to bet- ter understand player performance, improve team strategies, and reduce the risk of injuries. Coaches and analysts now use data to track player movements, predict match outcomes, and make smarter decisions. However, most of these advanced systems are expensive and are only available to professional clubs. Am- ateur and semi-professional players often dont have access to such tools because they require costly equipment and expert support. This gap shows the need for a simple, affordable plat- form that helps football players improve their skills using tools they already have, like smart- phones and basic video recordings. The project, AI-Driven Sports Analytics and Wellness Plat- form for Football Players, focuses on creating an easy-to-use digital system that gives play- ers access to professional-style analysis without needing expensive infrastructure. The platform combines video analysis, virtual coaching, and basic health tracking in one mobile application. The main goal of this system is to help play- ers understand their performance in a clear and practical way. It automatically analyzes match videos, tracks player movements, and creates a performance report similar to a school report card. This report includes useful details like running speed, stamina, passing accuracy, and overall consistency. In addition to game perfor- mance, the platform also supports player well- being by tracking simple wellness factors like sleep, nutrition, hydration, and recovery. The platform also includes a virtual AI coach that offers helpful suggestions, such as the best play-

    ing position for a player, customized training plans, and motivational support. The aim is to make every player feel like they have a personal mentor guiding them. To make the experience more engaging, the platform allows players to share their performance profiles, connect with coaches, and get noticed by scouts. Coaches can view player data, track progress over time, and give more personalized feedback. By com- bining performance tracking, wellness support, and social networking, the platform creates a supportive community where players can learn, grow, and feel motivated.

    1. Background

      Sports analytics has come a long way from sim- ple number-based tracking to more advanced technologies like machine learning and com- puter vision that can deliver real-time insights. In todays football environment, data plays a major role in planning team tactics, identify- ing talented players, and improving overall per- formance. However, while professional foot- ball clubs can afford costly sensors and ex- pert analysts, local and aspiring players often still depend on basic observation and traditional coaching methods. Because of this gap, there is a strong need for affordable and easy-to-use platforms that make modern sports analytics available to everyone. This project responds to that need by bringing together AI-powered per- formance analysis, virtual coaching, and well- ness monitoring into a single, user-friendly sys- tem. The goal is to create an accurate and acces- sible solution that supports not only athletic per- formance, but also the overall health and long- term development of football players.

    2. Relevance

      The proposed AI-driven sports analytics and wellness platform is highly relevant in todays fast-changing sports world. Football is be- coming increasingly popular at both grassroots and professional levels in India and around the world, and players now expect better ways to understand and improve their performance. Un- fortunately, most advanced analysis and coach- ing tools are still limited to elite clubs that can afford expensive equipment and expert support.

      As a result, many young and aspiring players miss out on valuable feedback and opportunities to grow. This project aims to reduce that gap by offering a low-cost, AI-powered platform that brings performance analysis, wellness tracking, and virtual coaching together in one easy-to-use system. It helps individual players track their progress, while also giving coaches, scouts, and sports academies reliable data to support fair and informed decisions. By adding social fea- tures and talent-sharing tools within the same application, the platform encourages interac- tion, visibility, and motivation among players. This creates a supportive and connected com- munity that can strengthen the future of football development in a sustainable way.

  2. Literature Review

    Recent advancements in football analytics in- creasingly leverage deep learning and com- puter vision to automate event detection, player tracking, and tactical understanding from match videos. Comprehensive surveys such as [1] highlight the evolution from tra- ditional CNNs to advanced architectures in- cluding 3D-CNNs, two-stream networks, and hybrid CNNRNN/LSTM models for spatio- temporal analysis of football events. These models enable fine-grained tasks such as ac- tion recognition, player identification, and se- mantic segmentation, while evaluation metrics like IoU, mAP, and MOTA ensure robust per- formance benchmarking. Practical implemen- tations discussed in applied works [4] further demonstrate the feasibility of deploying YOLO- based detection, tracking pipelines, and LSTM- driven event segmentation on single-camera football footage. However, both studies empha- ize challenges related to annotation scarcity, occlusion, camera perspective variations, and real-time inference latencygaps that moti- vate our platforms focus on lightweight mod- els, optimized preprocessing, and scalable de- ployment. Advanced spatial modeling ap- proaches, such as perspective-transform-based YOLO with weighted intersect fusion [5], fur- ther contribute to tactical insights by mapping detections onto standardized pitch layouts, in- fluencing our possession-forecasting and spatial performance metrics.

    Beyond vision-centric analytics, several studies focus on athlete health, performance op- timization, and talent evaluation using machine learning and AI-driven recommendation sys- tems. Research on ML-based performance and injury prediction [2] demonstrates the effective- ness of interpretable models like XGBoost in extracting actionable insights from physiologi- cal and workload data, while emphasizing real- time feedback for coaches and medical teams. Privacy-preserving architectures such as Split Federated Learning combined with RNNs [3] address scalability and data-security concerns in wearable-based health monitoring, directly informing the edge-cloud design of our Health and Wellness module. Unsupervised clustering techniques for football analysis [6] reduce de- pendency on labeled datasets by using color- segmentation and lighting adaptation, improv- ing robustness under real-world stadium con- ditions. Additionally, AI-driven recommenda- tion systems for nutrition [7] and player scout- ing [8] illustrate how generative AI and opti- mization techniques can enhance personalized dietary planning and talent identification. Col- lectively, these works shape our integrated plat- form by combining video analytics, health mon- itoring, nutrition guidance, and AI-based player profiling into a unified, coach-friendly decision- support system. Table. 1 shows the Compar- ative Analysis of Related Works in AI-Based Football Analytics

  3. Methods

    The proposed system follows a modular AI- driven framework designed to support compre- hensive football player development through four integrated functional components: AI Coaching, Player Performance Card Genera- tion, Health and Wellness Monitoring, and So- cial MediaBased Talent Showcasing. The methodology combines computer vision, ma- chine learning models, and user-centric mo- bile technologies to enable real-time player an- alytics, personalized feedback, holistic health tracking, and digital visibility. This struc- tured, multi-module approach ensures scal- ability, adaptability, and accessibility, mak- ing the platform suitable for grassroots, semi- professional, and professional football environ-

    ments.

    1. Video Acquistion and Preprocessing

      The system accepts match or training footage captured using consumer devices (smartphones, handheld cameras) or publicly available broad- cast/league videos. Because source videos vary in frame rate, resolution, viewpoint, and light- ing, a preprocessing pipeline standardizes and conditions input to improve downstream de- tection and tracking. The pipeline performs frame extraction, resolution and coordinate nor- malization, background/prior removal via se- mantic segmentation and illumination correc- tion. Fig. 1 shows vertical representation of the video acquisition and preprocessing pipeline. The pipeline standardizes heterogeneous foot- ball videos through frame extraction, normal- ization, background removal, and illumination correction prior to detection and tracking.

      Raw Video Input (Smartphone / Broadcast)
      Frame Extraction

      Resolution & Coordinate Normalization

      Background Removal (Semantic Segmentation)
      Illumination Correction
      Preprocessed Video Frames

      Figure 1: Vertical representation pipeline

    2. Player Detection and Training

      The keypoint detection is implemented using the YOLOv8-Pose model trained on a custom football keypoint dataset hosted on Roboflow.

      Table 1: Comparative Analysis of Survey

      Paper Techniques Used Core Area Strengths Limitations
      Review of DL

      Architectures for Football Video Analysis

      CNN, 3D-CNN,

      Two-stream Net- works, RNN, LSTM

      Player detection,

      tracking, event recognition

      Broad task-to-model

      mapping, metric analysis (mAP, IoU, MOTA)

      Survey-level

      study, no real-time im- plementation

      AI for Perfor-

      mance Enhance- ment and Injury Prevention

      XGBoost, Feature

      Engineering, Wear- able Data

      Performance ana-

      lytics, injury risk prediction

      Interpretable models,

      actionable insights for coaches

      Small datasets,

      limited vision- based analysis

      SFL-RNN Based

      Real-Time Health Monitoring

      Split Federated

      Learning, RNN, Wearables

      Athlete health

      monitoring, injury prediction

      Privacy-preserving,

      low latency, edge- cloud balance

      Network de-

      pendency, communica- tion overhead

      Advancing Foot-

      ball Game Analy- sis using CV and DL

      YOLO, Faster

      R-CNN, 3D-CNN, LSTM

      Player tracking,

      ball detection, event analysis

      Practical implementa-

      tion, real-time capa- bility

      Single-camera

      limitation, heuristic pre- processing

      Perspective Trans-

      form Based YOLO with WIF

      YOLO, Perspective

      Transform, Geo- metric Tracking

      Possession predic-

      tion, spatial analy- sis

      Tactical mapping, pos-

      session forecasting

      Camera

      calibration sensitivity, oc- clusion issues

      Unsupervised

      Clustering in Foot- ball Analysis

      Color Segmen-

      tation, K-means, Instance Segmenta- tion

      Team identifi-

      cation, lighting adaptation

      No labeled data re-

      quired, lighting ro- bustness

      Fails with

      similar jerseys, no player-level metrics

      FR-RANC:

      Nutrition-Centric Recommendation System

      Optimization Al-

      gorithms, AI-based Recommendation

      Personalized nutri-

      tion planning

      Region-aware, adapt-

      able dietary guidance

      Not athlete-

      specific, lacks sports perfor- mance focus

      FPSRec: Football

      Player Scouting System

      Machine Learning,

      Generative AI, Clustering

      Talent scouting,

      player recommen- dation

      Automated reports,

      human-readable in- sights

      No real-time

      video perfor- mance analysis

      First, the environment is initialized by installing the required libraries (ultralytics, roboflow) and verifying GPU availability (nvidia-smi). The code authenticates the Roboflow workspace through an API key and downloads the anno- tated football pose dataset in YOLO format. During training, YOLO computes pose loss,

      cal performance indicators in football analyt- ics. In our system, a YOLOv8-Pose model is trained using a custom football dataset obtained from Roboflow, containing manually annotated skeletal keypoints for players in diverse match environments. The dataset is represented as:

      box loss, and confidence loss while generating

      visual logs like F1 curves, confusion matrices, and training/validation batch predictions.

      Keypoint Detection Using YOLO-Pose us-

      N represents the total number of training samples (images) in your dataset. Each input image Ii is associated with a structured keypoint

      ing Accurate pose estimation is essential for analyzing player movements, identifying

      set

      joint dynamics, and extracting biomechani- where vi,k indicates the visibility of each joint

      (1 = visible, 0 = occluded).

      During training, YOLO-Pose jointly opti- mizes bounding-box prediction and keypoint localization. The bounding-box quality is mea- sured using Intersection-over-Union, defined as:

      footage, a YOLOv8-X object detection model is trained on a custom football dataset obtained from Roboflow. The dataset contains annotated player bounding boxes captured from diverse match scenarios.

      nates, objectness confidence, and class prob- abilities simultaneously. The bounding-box

      • Bpred : Predicted bounding box generated by the YOLO-Pose model.
      • Bgt : Ground-truth bounding box provided in

        regression quality is optimized using the

        Complete-IoU (CIoU) loss, computed as:

        the annotated dataset.

      • Bpred Bgt : Intersection area between the

        predicted box and the ground-truth box.

      • Bpred Bgt : Union area covering both the predicted and ground-truth boxes.
      • IoU : Intersection-over-Union metric used to evaluate bounding-box localization accuracy. which provides an overlap score used to penal-

        ize inaccurate player localization.

        The overall multi-task objective for pose

        where IoU is the Intersection-over-Union between predicted and ground-truth boxes, (·)

        is the Euclidean distance between box centers, c is the diagonal length of the smallest enclosing box, v is an aspect-ratio consistency term, and is a balancing factor.

        The IoU component is given by:

        learning integrates bounding-box loss, keypoint loss, object presence confidence, and class probability, expressed as:

      • Ltotal : Overall multi-task loss optimized by the YOLOv8-Pose model.
      • Lbox : Bounding-box regression loss, measur- ing the localization error between predicted and ground-truth boxes.
      • Lkp : Keypoint regression loss computed us- ing Smooth L1 over all visible joints.
      • Lobj : Objectness loss indicating confi- dence of player presence within the predicted bounding box.
      • Lcls : Classification loss for predicting the correct class label (player or background).
      • box : Weighting coefficient controlling the contribution of bounding-box loss.
      • kp : Weighting coefficient for keypoint re- gression loss.

        YOLOv8 incorporates an objectness predic-

        tion branch, trained using binary cross-entropy:

        where yi {0, 1} is the ground-truth object- ness label and yi is the predicted confidence.

        Class probabilities are optimized using a standard cross-entropy loss:

        where C is the number of classes, pc is the true class label, and pc is the predicted proba- bility.

        During training, the model is optimized for 50 epochs using the yolov8x.pt backbone,

      • obj

        : Weight factor for the objectness confi-

        a batch size of 6, and an image resolution

        of 1280 × 1280, ensuring high spatial fidelity

        dence term.

      • cls : Weight factor for the classification term. To accurately localize players across match

      for long-range player detection. A heatmap is generated by aggregating player position coor- dinates across all video frames and mapping

      the frequency of occurrence onto a normalized football field grid. Regions with higher ac- tivity density are visualized using warmer col- ors, enabling intuitive analysis of player move- ment patterns and tactical behavior. Player positions are extracted from consecutive video frames using the YOLOv8-based detection and tracking module. These coordinates are accu- mulated over time and projected onto a nor- malized football field representation. Regions with higher player presence are highlighted us- ing warmer colors, while low-activity regions are represented with cooler tones. This visu- alization provides intuitive insights into player movement patterns, positional discipline, and tactical behavior, supporting both performance evaluation and strategic decision-making. As shown in Fig. 2, the generated heatmap reveals dominant player activity zones and positional trends throughout the match.

      Figure 2: HeatMap Generated

    3. Performance extraction

      To generate the Player Performance Card, the system extracts a comprehensive set of spatio- temporal performance metrics from tracked player trajectories across video frames. Us- ing continuous player tracking, motion-based features such as total distance covered, aver- age speed, maximum speed, sprint count, ac- celeration and deceleration events are com- puted to quantify physical workload and inten- sity. Speed thresholds are applied to distinguish high-intensity runs and sprints, while frame-to- frame displacement is used to estimate accel- eration dynamics. Positional context is inferred by mapping player coordinates onto field zones, enabling role-based insights such as defensive, midfield, or attacking dominance. These raw metrics are normalized across match duration

      and aggregated into interpretable performance indicators, which are visually summarized in the player card along with positional labels and tracking consistency. This structured repre- sentation enables objective player comparison, workload monitoring, and data-driven perfor- mance assessment for coaches and analysts.

    4. Player Card Generation-Module 1

      The Player Performance Card presents the ex- tracted metrics in a structured and interpretable format, similar to a performance report used by professional coaches and analysts. Each raw feature (speed, stamina, accuracy, and consis- tency) is normalized to ensure fairness across sessions and players, and then aggregated into intuitive performance indices.

      The card summarizes multiple dimensions of a players capabilities, including physical fit- ness indicators, technical skill metrics (such as pass or shot effectiveness), and behavioral con- sistency across matches. A weighted scoring mechanism combines these indices to produce an overall performance rating that reflects the players strengths and areas for improvement. The card is dynamically updated after each ses- sion and stored within the players profile, en- abling long-term trend analysis, comparative evaluations, and easy sharing with coaches or scouts.

      Fig.3 illustrates the Player Performance Card generated by the proposed AI-based foot- ball analytics system. The interface enables dy- namic player selection from detected on-field players, each uniquely identified and color- coded for clear tracking throughout the match. Once a player is selected, the system aggre- gates spatio-temporal data extracted from video frames to generate an individualized perfor- mance summary.

      The player card presents key physical per- frmance metrics, including total distance cov- ered, average speed, and maximum speed achieved during the session. These metrics are computed using player trajectory tracking and frame-to-frame motion analysis. The associated player snapshot provides visual confirmation of the tracked identity, improving interpretability and trust in the extracted data.

      ity multiplier reflects the athletes training in- tensity.

      Figure 3: Player Card Generated

    5. Health and Wellness Monitoring- Module 2

      The Health and Wellness Monitoring mod- ule provides a comprehensive framework for assessing athlete health, metabolic condition, dietary intake, and recovery patterns. This module integrates physiological computations, lifestyle tracking, and machine learning based analysis to support continuous monitoring of player wellness. The backend implementa- tion is developed using the Flask framework, enabling athletes to input relevant health at- tributes such as height, weight, age, sleep du- ration, hydration level, activity level, and fit- ness goals. These attributes are stored in a cen- tralized database to enable longitudinal tracking and personalized recommendations.

      1. Metabolic Rate Estimation

        To estimate the athletes baseline metabolic requirement, the system computes the Basal Metabolic Rate (BMR) using the MifflinSt Jeor equation, which is widely accepted for es- timating resting energy expenditure.

        where W represents body weight in kilo- grams, H represents height in centimeters, and A represents age in years.

        The calculated BMR is adjusted based on the athletes activity level to estimate the Total Daily Energy Expenditure (TDEE). The activ-

        TDEE = BMR × ActivityFactor (11)

        Typical activity multipliers include seden- tary (1.2), moderately active (1.55), and highly active (1.9). Based on the calculated TDEE, the system determines an appropriate caloric intake recommendation depending on the ath- letes goal such as weight maintenance, muscle gain, or fat reduction.

      2. Body Composition Analysis

        To evaluate general body composition, the sys- tem calculates the Body Mass Index (BMI). Al- though BMI alone does not fully represent ath- letic fitness, it provides a useful baseline indica- tor when combined with other metabolic mea- surements.

        where W represents body weight in kilo- grams and H represents height in meters.

        Based on BMI values, the system catego- rizes body composition into underweight, nor- mal weight, overweight, or obese ranges, which helps identify potential health concerns affect- ing athletic performance.

      3. Dietary Image Analysis using Deep Learning

        To analyze nutritional intake, the module in- corporates an image based machine learning pipeline. Athletes can upload images of food items which are processed by a deep learn- ing model implemented using TensorFlow and Keras.

        Before model inference, the input image un- dergoes several preprocessing steps including resizing to a standardized resolution, normal- ization of pixel values, and conversion into a tensor format suitable for neural network pro- cessing.

        where X represents the original pixel inten- sity and Xnorm represents the normalized pixel value.

        The processed image is then passed through a convolutional neural network (CNN) architec- ture trained for food classification. CNN mod- els are particularly effective for visual recogni- tion tasks due to their ability to capture spatial patterns and hierarchical features from images. The convolution operation applied in CNN lay- ers is defined as:

        ommendation process integrates metabolic re- sults, food analysis outcomes, and lifestyle in- puts such as sleep duration and hydration level. For example, if caloric intake is significantly lower than the estimated TDEE, the system rec- ommends increased nutritional intake to main- tain optimal energy levels. Similarly, insuffi- cient sleep patterns or excessive workload indi- cators trigger recovery recommendations such as reduced training intensity or increased rest

        intervals.

        [1] Collect user wellness inputs (height, weight, age, sleep, hydration) Compute BMR using MifflinSt Jeor equation Estimate TDEE

        m n using activity multiplier Calculate BMI for

        where I represents the input image and K

        represents the convolution kernel.

        The CNN architecture typically includes multiple convolution layers followed by pool- ing layers for feature extraction, fully connected layers for classification, and a Softmax output layer that produces the probability distribution over food classes.

        body composition analysis Receive uploaded food image Preprocess image and normalize pixel values Apply CNN model to classify food item Retrieve nutritional information from food dataset Compare caloric intake with recom- mended TDEE Generate dietary and wellness recommendations

        By integrating physiological computations, deep learning based dietary recognition, and structured nutritional data analysis, the Health

        and Wellness Monitoring module provides a

        where zi represents the output score for class

        i and K represents the total number of classes.

      4. Nutritional Dataset Integration

        The predicted food item from the deep learn- ing model is matched against a structured nutri- tional dataset containing more than 1000 food items from multiple cuisines including Indian, Italian, Chinese, Mexican, Japanese, Thai, and American dishes. Each food entry contains de- tailed attributes such as caloric value, protein content, carbohydrate levels, fat composition, fiber content, preparation time, and dietary suit- ability for athletes.

        This dataset allows the system to compute the estimated caloric intake from the detected food item and compare it with the athletes rec- ommended daily energy requirement.

      5. Wellness Recommendation Generation

        After calculating metabolic indicators and nu- tritional intake, the system generates person- alized wellness recommendations. The rec-

        comprehensive framework for tracking athlete health and supporting performance optimiza- tion.

    6. Injury Prevention-Module 3

      The Injury Risk Prediction and Recovery Mod- ule is designed to proactively assess the likeli- hood of sports-related injuries and recommend suitable recovery strategies based on player workload and physical stress indicators. Unlike traditional wearable-based monitoring systems, this module relies on performance metrics ex- tracted from video analysis and historical work- load patterns to evaluate injury susceptibility.

      The module analyzes key match-related pa- rameters such as sprint count, total distance covered, acceleration and deceleration fre- quency, maximum running speed, and high- intensity activity duration. These indicators collectively reflect the biomechanical load and muscular stress experienced by a player dur- ing a match. Based on these inputs, a compre- hensive risk score is computed to represent the overall injury likelihood.

      Players are categorized into multiple injury risk levels ranging from very low to critical. This classification enables early identification of high-risk players who may require imme- diate rest or medical attention. Additionally, position-specific injury tendencies are incorpo- rated to improve prediction reliability, as dif- ferent playing roles exhibit distinct physical de- mands and injury patterns.

      The system further predicts the most prob- able injury types, such as hamstring strain, groin strain, kee ligament stress, calf strain, or muscle fatigue. These predictions are derived by correlating workload indicators with known football injury patterns. The output includes injury severity estimation and contributing risk factors to enhance interpretability.

      To complement injury prediction, a fatigue assessment mechanism evaluates cumulative physical exhaustion using a fatigue score. This score reflects both acute and prolonged work- load effects, enabling the system to distinguish between normal post-match fatigue and exces- sive physical strain.

      Based on the injury risk level and fatigue as- sessment, the module generates a personalized recovery plan. The recovery recommendations include rest duration, hydration requirements, sleep guidelines, nutrition support, and suitable recovery activities such as stretching, ice ther- apy, and active recovery sessions. In high-risk cases, medical consultation warnings are also issued.

      [1]

      Player match statistics: sprint count S, to- tal distance TD, accelerations A, decelerations D, maximum speed V , high-intensity distance HID, player position P

      Injury risk score, predicted injuries, and re- covery plan

      Collect player performance metrics from video analysis

      Compute injury risk score R using workload indicators

      Apply biomechanical modifiers based on high speed and acceleration

      Apply position-based multiplier according to player role

      Normalize the risk score to obtain final in- jury risk value

      Classify the player into one of the eight in- jury risk levels

      Predict probable injury types based on work- load patterns

      Calculate fatigue score F using sprint count, distance, accelerations, decelerations, and high- intensity running

      Determine fatigue level using predefined fa- tigue thresholds

      Generate recovery plan including:

      • Rest duration
      • Hydration guidelines
      • Sleep recommendations
      • Recovery activities
      • Nutrition suggestions

      Return injury risk assessment and personal- ized recovery plan

      Overall, the Injury Risk Prediction and Recovery Module supports preventive sports healthcare by enabling informed decision- making for training load management, recov- ery planning, and injury mitigation. This con- tributes to improved player safety, performance sustainability, and long-term athletic develop- ment.

    7. AI Coaching and Recommendation Engine-Module 4

      The AI Coaching and Recommendation Engine functions as the intelligent decision-making layer of the system, transforming raw perfor- mance metrics and wellness indicators into ac- tionable training insights for athletes. This module combines rule-based evaluation strate- gies with machine-learningassisted analysis to generate personalized coaching recommenda- tions tailored to each players performance pro- file. The engine processes multiple input fea- tures collected from the performance analyt- ics and wellness monitoring modules. These inputs include quantitative performance indi- cators such as speed scores, agility metrics, passing accuracy, stamina levels, and decision- making efficiency. Each metric is represented as a normalized performance score ranging from 0 to 100, allowing the system to com- pare attributes across different skill dimensions. Based on these scores, the module identifies key strengths and areas requiring improvement.

      For example, attributes such as Speed Agility, Passing Accuracy, and Decision Making are evaluated as strength indicators when their per- formance values exceed predefined thresholds. Conversely, metrics such as Finishing efficiency or stamina levels are flagged as improvement areas when their values fall below the ex- pected benchmark. The system then associates these weaknesses with specific corrective train- ing drills and conditioning exercises. In ad- dition to skill analysis, the AI coaching mod- ule also performs positional suitability assess- ment. By analyzing a combination of speed metrics, passing accuracy, tactical awareness indicators, and endurance levels, the system es- timates the most suitable playing position for the athlete. The predicted position is accom- panied by a confidence score, which represents the systems certainty in the recommendation based on the evaluated feature set. Another key function of the coaching engine is the genera- tion of personalized weekly training plans. The system dynamically constructs training sched- ules by mapping identified weaknesses to rel- evant drills and conditioning exercises. For instance, if finishing accuracy is identified as a weakness, the training module may recom- mend drills such as one-on-one finishing prac- tice, set-piece training, and small-sided pos- session games. Similarly, low stamina indica- tors trigger endurance-focused exercises such as sprint intervals, shuttle runs, and agility lad- der drills. The training schedule is organized into structured weekly sessions, each focusing on specific skill areas such as finishing drills, high-intensity conditioning, tactical awareness, or match preparation. This structured approach ensures that athletes receive balanced develop- ment across technical, physical, and tactical as- pects of the game. To provide additional mo- tivation and feedback, the system also gener- ates contextual coaching insights based on ob- served performance trends. These insights sum- marize recent improvements and highlight ar- eas requiring sustained effort. For instance, if a players speed metrics show measurable im- provement across multiple sessions, the sys- tem communicates this progress through moti- vational feedback messages while encouraging continued focus on weaker attributes. By con-

      tinuously analyzing performance trends, iden- tifying strengths and weaknesses, and gener- ating adaptive training plans, the AI Coach- ing and Recommendation Engine enables data- driven athlete development. This approach al- lows players and coaches to make informed training decisions, optimize skill progression, and reduce the risk of performance stagnation or overtraining.

      1. Performance Evaluation Model

        To quantify an athletes overall performance level, the system computes a composite per- formance score using multiple skill indicators. Each skill metric is normalized to a range be- tween 0 and 1 before aggregation.

        where P represents the overall performance score, Si represents the normalized value of the ith skill metric, and wi represents the weight as- signed to that metric. The weights are deter- mined based on the importance of each skill in football performance analysis.

        This aggregated score helps the system iden- tify strengths and weaknesses by comparing in- dividual skill metrics with predefined bench- mark thresholds.

      2. Personalized Fitness Plan Generation

        The Personalized Fitness Plan Generation mod- ule is designed to recommend an appropriate workout and nutrition plan based on the users fitness profile. The system collects several inputs including fitness level, workout goals, available equipment, workout frequency, ses- sion duration, weight goal, and existing med- ical conditions. These parameters are used to generate a customized training and dietary rec- ommendation.

        Initially, the system gathers user input through a fitness questionnaire interface. Mandatory fields such as fitness level, workout goals, workout days, and session duration are validated before processing. Based on the provided fitness level, the system selects an appropriate exercise dataset consisting of

        beginner, intermediate, or advanced exercises.

        The system then constructs a structured workout schedule by iterating through the se- lected workout days and assigning exercises ac- cording to the users goals and available equi- ment. The overall procedure followed by the system is described in Algorithm 3.7.2.

        Fitness Level L, Goals G, Equipment E, Workout Days D, Session Duration T , Weight Goal W , Health Conditions H

        Personalized Workout Plan and Meal Plan Collect user fitness parameters Validate

        mandatory inputs Select exercise dataset based on fitness level L

        day = 1 to D Initialize workout schedule for the day goal includes strength or muscle gain Select strength exercises compatible with equipment E goal includes weight loss or en- durance Add suitable cardio exercises goal in- cludes flexibility Include stretching routines

        Generate meal plan based on weight goal W Modify nutrition recommendations accord- ing to health conditions H Initialize progress tracking

        Workout plan and nutrition plan

        After generating the workout schedule, the system produces a complementary meal plan tailored to the users weight management goal. The nutrition plan provides dietary suggestions that help support muscle gain, weight loss, or general fitness improvement. Additionally, health conditions specified by the user are con- sidered when recommending foods to avoid po- tential health risks.

        Finally, the system initializes a progress tracking component that enables users to moni- tor their workout completion and maintain con- sistency in their training routine. This inte- grated approach ensures that the generated fit- ness plan is personalized, structured, and adapt- able to the users health profile and fitness ob- jectives.

      3. Input and Output Description

        The inputs to the Personalized Fitness Plan Generation module consist of user-provided fit- ness parameters collected through the fitness questionnaire interface. These inputs include fitness level, workout goals, available equip-

        ment, number of workout days per week, pre- ferred workout duration, weight management goal, and any existing medical conditions.

        Based on these inputs, the system processes the data using the algorithm described above to generate a structured output consisting of a per- sonalized workout schedule, recommended ex- ercises for each training day, and a complemen- tary nutrition plan. The system also initializes a progress tracking mechanism that allows users to monitor their workout completion and main- tain consistency in their training routine.

    8. Social Talent Promotion(Entire App)

      The Social Talent Promotion module enhances player visibility and engagement by integrating a controlled social-sharing ecosystem within the platform. This module allows athletes to publish their performance cards, match high- lights, and key statistics to a dedicated football- centric network. Unlike conventional social platforms, this system is designed specifically for athletic development, enabling players to showcase verifiable, data-driven performance outputs generated directly from the analytics pipeline.

      Each shared post is linked to authenticated player data and includes optional highlight clips, detected events, or session summaries produced by the systems video-analysis mod- ules. Coaches, scouts, and academies can browse player profiles through search and rec- ommendation functionalities. These recom- mendations are generated using similarity anal- ysis based on skill vectors, movement features, and historical performance indices, enabling talent discovery even at lower or amateur lev- els.

      fig. 4 illustrates a sample visualization from the proposed AI-driven football analytics and social engagement module. The upper part of the figure shows a football-related image repre- senting raw visual input captured from match or training footage using consumer devices such as smartphones or handheld cameras. This vi- sual input forms the basis for downstream com- puter vision processing, including player detec- tion, activity recognition, and contextual analy- sis.

      This operates through a structured interaction

    9. System Architecture

      The proposed system operates through a multi- layer pipeline that begins with data acquisi- tion from match videos and wellness inputs up- loaded via a mobile application. The AI pro- cessing layer performs player detection, track- ing, and pose estimation to extract speed, dis- tance, accuracy, and stamina-related metrics. These outputs feed into the application logic layer, where wellness analytics and AI-driven coaching recommendations are generated. Fi- nally, the presentation layer delivers person- alized dashboards, health insights, and social- networkbased features to players and coaches in an intuitive interface.

      Data Acquisition Layer

      Mobile Application (Player/Coach Upload)

      Match Videos & Wellness Details

      Performance Metric Extraction (Speed, Accuracy, Stamina)

      Figure 4: Sample visualization

      AI Processing Layer

      Player Detection YOLOv8, ByteTrack

      Appearance ReID Performance Metric

      Extraction

      pipeline that connects the mobile application, backend API, database, and media storage ser- vices. When a user logs into the applica- tion, authentication is performed using JSON

      Application Logic Layer

      AI Coaching Engine Wellness Analytics Recommendation Engine

      Presentation Layer

      Injury & Recovery Analysis

      Web Tokens (JWT). Upon successful authen- tication, the backend generates a secure to- ken that allows the user to access protected re- sources within the system.When a user creates a post containing text, images, or videos, the media files are uploaded to Cloudinary for ef- ficient storage and delivery. The backend then stores the post metadata and references to the media files in the MongoDB database. Other users can interact with the post through likes, comments, and shares. These interactions are processed by the backend APIs and updated in real time.Player profile data such as statis- tics, physical attributes, and achievements are stored in the database and dynamically rendered on the user interface, creating a LinkedIn-style sports profile. This allows players to showcase their performance data, achievements, and me- dia highlights to other users within the platform.

      Performance Dashboard Wellness Insights Recovery Status Social Networking

      Figure 5: System Architecture

  4. Experiments and Results

    This section presents a comprehensive ex- perimental evaluation of the proposed AI- driven football analytics and wellness monitor- ing framework. The experiments are designed to assess the effectiveness, robustness, and real- time feasibility of the system across its core modules, including player detection, pose esti- mation, performance metric extraction, and vi- sualization. Quantitative results and qualita- tive analyses are reported using diverse football video datasets to demonstrate the systems ac- curacy, generalization capability, and practical applicability under varying match conditions.

    1. Datasets

      To evaluate the proposed AI-driven football an- alytics framework, experiments were conducted on multiple football video datasets collected from Kaggle, Scoutingfeed and supplementary match recordings. These datasets include di- verse match scenarios with annotated on-field entities such as players, goalkeepers, and ref- erees, enabling reliable evaluation under real- world conditions. Figure 6 illustrates represen- tative sample frames from the football video datasets used for training and evaluation, high- lighting player detection and match diversity and referee annotations under different match conditions.

      Figure 6: Sample frames from datasets

      1. Football Field Detection Dataset

        The football field detection dataset consists of 2,847 images focusing on field geometry and keypoint localization. The dataset was split into 1,992 training images (70%), 570 valida-

        tion image (20%), and 285 test images (10%). Each image is annotated with 29 keypoints rep- resenting essential field elements such as cor- ner flags, penalty boxes, center circle, and goal posts. All images are provided at a resolution

        of 1920×1080 pixels.

      2. Football Player Detection Dataset

        The football player detection dataset contains 4,521 images annotated for multi-class player detection. The dataset was divided into 3,165 training images (70%), 904 validation images

        (20%), and 452 test images (10%). On average, each image contains 18.3 players. The anno- tated classes include home team players, away team players, referees, and goalkeepers. The

        image resolution for this dataset is 1280×720 pixels.

    2. Training Performance
      1. Football Field Keypoint Detection

        The YOLOv8x-pose model was trained for 100 epochs using an NVIDIA Tesla V100 GPU. The model achieved stable convergence with a fi- nal training loss of 0.0187, indicating effec- tive learning of both bounding box localization and keypoint regression. Convergence was ob- served at epoch 73, with minimal improvement in subsequent epochs.

        The validation results demonstrate strong detection and localization performance, achiev- ing a box mAP@0.5 of 96.2% and a keypoint mAP@0.5 of 93.5%. The optimal F1-score of 0.933 was obtained at a confidence threshold of 0.45, reflecting a balanced precisionrecall trade-off.

      2. Football Player Detection

        The YOLOv8x object detection model was trained for 50 epochs on the player detection dataset. The final training loss reached 0.0241, showing stable optimization across box re- gression, objectness, and classification compo- nents. Validation results indicate reliable multi- class detection performance, with an overall mAP@0.5 of 91.7% and recall of 87.1%.

        Home and away players achieved the highest detection accuracy due to their frequent appear- ance in training data, while referees and goal- keepers showed slightly lower performance due to class imbalance.

    3. Test Set Evaluation
      1. Field Detection Results

        The trained YOLOv8x-pose model was eval- uated on 285 unseen test images. The model achieved a keypoint localization accu- racy (PCK@0.1) of 94.2% and a field detection rate of 99.3%. Performance remained stable across different environmental conditions, in- cluding indoor settings, rainy scenes, and night matches.

        The average inference time was 47 ms per image, corresponding to a processing speed of

        21.3 frames per second, enabling near real-time field analysis.

      2. Player Detection Results

        Evaluation on 452 test images demonstrated strong generalization, achieving an mAP@0.5 of 90.4% and an average recall of 86.2%. De- tection accuracy decreased under heavy occlu- sion and long-range views but remained robust across most match scenarios.

        The model achieved a multi-object tracking accuracy (MOTA) of 84.7%. Inference speed averaged 68 ms per image, allowing real-time deployment in video-based analytics systems.

    4. Comparative Analysis

      The proposed YOLOv8-based framework was compared with existing baseline methods for both field detection and player detection. The results demonstrate that YOLOv8x-pose achieves superior accuracy while maintaining higher inference speed compared to previous YOLO versions and two-stage detectors.

      Table 2: Comparison of Football Field Detec- tion Methods

      data augmentation on detection performance and computational efficiency.

          1. Impact of Transfer Learning

            We evaluated three training strategies: ran- dom initialization, ImageNet pre-training, and COCO pre-training. The results are presented in Table 4.

            The results demonstrate that transfer learn- ing from the COCO dataset significantly ac- celerates model convergence while improving detection accuracy. Compared to random ini- tialization, COCO pre-training reduced training time by approximately 66.4% and increased ac- curacy by 14.9%. This highlights the effective- ness of leveraging large-scale object detection knowledge for football scene understanding.

          2. Impact of Image Resolution

            We analyzed the effect of different input image resolutions on detection accuracy, inference la- tency, and memory consumption. The results are summarized in Table 5.

            While higher resolutions improve detection accuracy by preserving fine-grained spatial de- tails, they introduce substantial computational

            Method mAP@0.5 PCK@0.1 FPS overhead. The 640×640 resolution provides an

            Faster R-CNN 88.3% 89.7% 8.2
            YOLOv5x-pose 91.2% 92.4% 18.6
            YOLOv7-pose 93.1% 93.8% 19.3

             

            optimal trade-off between accuracy and real-

            time performance, making it suitable for live video analysis.

            YOLOv8x-pose 96.2% 94.2% 21.3

            Table 3: Comparison of Football Player Detec- tion Methods

    5. Ablation Studies

      To better understand the contribution of indi- vidual design choices and training strategies, we conducted a series of ablation experiments. These studies analyze the impact of transfer learning, input image resolution, and mosaic

          1. Impact of Mosaic Augmentation

      Mosaic data augmentation was evaluated to as- sess its influence on both field detection and player detection tasks.

      Disabling mosaic augmentation significantly improved field detection accuracy by 6.5% by preserving global geometric relationships criti- cal for keypoint localization. However, player detection accuracy slightly decreased by 1.5%, confirming that mosaic augmentation benefits object diversity but disrupts spatial consistency required for field geometry estimation.

    6. Error Analysis

      Despite strong overall performance, several failure cases were identified.

      For field detection, errors primarily occurred

      Table 4: Impact of Transfer Learning on Field Detection

      Training Strategy mAP@0.5 Training Time
      Random Initialization 81.3% 42.6 hours
      ImageNet Pre-training 88.7% 18.4 hours
      COCO Pre-training (Ours) 96.2% 14.3 hours

      Table 5: Impact of Image Resolution on Performance

      Resolution mAP@0.5 Inference Time Memory Usage
      416×416 84.2% 28 ms 2.1 GB
      640×640 91.8% 47 ms 3.8 GB
      1280×1280 95.4% 68 ms 7.2 GB
      1920×1920 96.1% 142 ms 12.4 GB

      under severe weather conditions such as fog and heavy rain, resulting in an 8.6% failure rate. Unusual camera angles, particularly aerial or top-down views, accounted for a 6.3% failure rate, while partially visible or cropped fields contributed to a 4.7% failure rate.

      For player detection, dense player lusters with more than five overlapping players led to a 12.4% miss rate. Similar jersey colors be- tween teams caused an 8.9% misclassification rate, and motion blur during fast player move- ments resulted in a 7.2% miss rate.

      These observations highlight the remaining challenges in extreme visual conditions and dense scenes.

    7. Training and Validation Learning Curves

      Figure 7 presents the training and validation learning curves of the YOLOv8-based detection models. The training losses, including bound- ing box loss, classification loss, and distribu- tion focal loss, exhibit a smooth and consistent decline throughout the training process, indi- cating stable optimization and effective feature learning. The absence of significant oscillations confirms that the learning rate and optimization strategy were well tuned.

      The validation loss curves closely follow the training loss trends, demonstrating strong gen- eralization capability and minimal overfitting. After a rapid reduction in the initial epochs, the validation losses stabilize, suggesting that the model converges effectively while maintaining

      robustness on unseen data.

      Precision and recall metrics increase sharply during the early epochs and gradually plateau after approximately 20 epochs, indicating that the model quickly learns to identify relevant objects and progressively improves detection completeness. The consistently high precision and recall values in later epochs reflect reliable detection performance.

      Similarly, the mAP@0.5 and mAP@0.5:0.95 metrics show continuous improvement with increasing epochs, with rapid gains during early training followed by gradual refinement. The convergence of these metrics confirms accurate object localization and confidence estimation across multiple IoU thresholds. Overall, the learning curves validate the effectiveness of the YOLOv8 architecture and training strategy for football field and player detection tasks. fig. 7 shows the training and validation learning curves of the YOLOv8 model showing loss convergence, precision, recall, and mean Average Precision (mAP) across epochs.

    8. Confusion Matrix Analysis

      Figure 8 presents the confusion matrix of the proposed YOLOv8-based player detection model evaluated on the test dataset. The ma- trix exhibits strong diagonal dominance, indi- cating high classification accuracy across all major categories. In particular, the player class achieves the highest number of correct predic- tions, reflecting the effectiveness of the model

      Figure 7: Training and validation learning curves

      Figure 8: Confusion matrix

      in identifying frequently occurring on-field en- tities.

      Minor misclassifications are observed be- tween visually similar classes such as goal- keepers and players, as well as referees and background regions. These errors are primar- ily caused by partial occlusions, similar jersey colors, and reduced spatial resolution for dis- tant players. Additionally, a small number of background regions are incorrectly classified as players in highly crowded scenes, highlighting the inherent challenges of dense football envi- ronments.

      Overall, the confusion matrix analysis

      confirms that the proposed detection frame- work maintains strong discriminative capability while remaining robust under real-world match conditions. The observed error patterns are con- sistent with the reported precision, recall, and F1-score metrics, further validating the reliabil- ity of the system for real-time football analyt- ics applications. Fig. 8 shows Confusion matrix of the YOLOv8-based player detection model on the test dataset. Rows represent ground- truth classes, while columns represent predicted classes for ball, goalkeeper, player, referee, and background categories.

    9. Real-World Application Performance

      To evaluate practical usability, the integrated system was deployed for live football match analysis. Performance metrics were recorded across 15 full match recordings.

      The system achieved an end-to-end process- ing latency of 115 ms per frame and a video analysis throughput of 8.7 FPS at a resolution of 1280×720. Field tracking consistency reached 97.8% across consecutive frames, while player

      tracking consistency achieved 91.4%. System uptime remained at 99.2% throughout the de- ployment.

      In total, the system processed 12.5 hours of match footage, detecting an average of 27.4 field keypoints and 19.3 players per frame with

      high temporal stability, demonstrating its suit- ability for real-world deployment.

    10. Discussion

      The experimental evaluation confirms that YOLOv8-based architectures achieve state-of- the-art performance for football field and player detection tasks. Transfer learning significantly accelerates convergence while improving ac- curacy, particularly when pre-trained on large- scale datasets such as COCO. Higher-resolution inputs enhance fine-grained keypoint localiza- tion, though at increased computational cost.

      The ablation results further demonstrate that preserving spatial integrity is essential for field keypoint detection, explaining the performance gains observed when mosaic augmentation is disabled. Overall, the proposed system achieves a strong balance between accuracy, robust- ness, and real-time performance, making it well suited for live broadcast analysis and advanced football analytics applications.

  5. Conclusion and Future Scope

    This work presented an AI-driven football an- alytics framework that integrates football field keypoint detection and multi-class player detec- tion using YOLOv8-based deep learning archi- tectures. Extensive experiments conducted on diverse football datasets demonstrate that the proposed system achieves high accuracy, ro- bust generalization, and real-time performance under varying illumination conditions, camera viewpoints, and match scenarios. The field key- point detection module effectively captures ge- ometric structures of the football pitch, while the player detection module reliably distin- guishes between players, goalkeepers, refer- ees, and background regions. Ablation studies further confirm the benefits of transfer learn- ing, optimized input resolution, and carefully selected data augmentation strategies. Over- all, the proposed framework proves to be well- suited for real-world football video analysis and provides a strong foundation for advanced sports analytics applications.

    Despite the strong performance of the pro- posed system, several extensions can further en- hance its capabilities. Future work will focus on

    integrating temporal modeling and multi-object tracking to enable continuous player identifica- tion and trajectory analysis across video frames. Incorporating player re-identification and team formation analysis can provide deeper tactical insights. Additionally, extending the frame- work to support ball tracking, event detection (such as passes, shots, and fouls), and player performance metrics will further enrich match analytics. Optimizing the models for edge de- ployment and low-latency inference can facil- itate real-time analysis for live broadcasts and stadium-based systems. Finally, expanding the dataset to include more diverse leagues, weather conditions, and camera setups will improve ro- bustness and scalability for global deployment.

  6. References
  1. A. M. Esa, N. Jamil, M. F. M. Mohsin, and M. A. M. Ali, A Re- view of Deep Learning Architectures for Automated Video Anal- ysis in Football Events, in Proc. 2023 3rd International Con- ference on Computer and Communication Engineering (ICCCE), 2023, pp. 16.
  2. T. D. Shukla, D. Nimma, K. S. Pokkuluri, S. Najmusaqib, K. K. Sivakumar, and B. K. Bala, Utilizing Artificial Intelligence for Enhancing Performance and reventing Injuries in Sports Analyt- ics, in Proc. 2024 International Conference on Intelligent Com- puting and Sustainable Innovations in Technology (IC-SIT), 2024,

    pp. 17.

  3. R. Ferdous, T. K. Das, K. Gupta, and D. Goswami, Real-Time Health Monitoring for Athletes in Dynamic Sports Environments Based on Split Federated Learning and Recurrent Neural Net- works, in Proc. 2024 International Conference on Emerging Technologies in Computing (ICETC), 2024, pp. 110.
  4. A. H. Benakesh and R. Rajeev, Advancing Football Game Anal- ysis: Integrating Computer Vision, Deep Learning, and Hybrid Techniques for Enhanced Video Analytics, in Proc. 2024 Inter- national Technology Conference on Smart Computing for Innova- tion and Advancement in Industry 4.0 (OTCON), 2024, pp. 110.
  5. Y. Author et al., Perspective Transform Based YOLO with Weighted Intersect Fusion for Forecasting the Possession Se- quence, in Proc. IEEE Conf. on Sports Analytics, 2023.
  6. X. Author et al., Unsupervised Clustering in Football Analy- sis: A Color-Segmentation and Lighting Adaptation Approach, in Proc. IEEE Conf. on Computer Vision Applications in Sports, 2022.
  7. M. Siddiqui, F. Akther, G. M. E. Rahman, and R. Mostafa, A Regionally Adaptable Nutrition-Centric Food Recommendation System (FR-RANC), IEEE Access, vol. 13, pp. 150179150183,

    2025.

  8. A. M. Rinaldi, A. Romano, C. Russo, and C. Tommasino, FP- SRec: Football Players Scouting Recommendation System Based on Generative AI, in Proc. IEEE Int. Conf. on Big Data, 2024,

    pp. 71417144.

  9. M. Siddiqui, F. Akther, G. M. E. Rahman, and R. Mostafa, A Regionally Adaptable Nutrition-Centric Food Recommendation System (FR-RANC), IEEE Access, vol. 13, pp. 150179150183, 2025.
  10. J. Li, Integrating Generative AI for Enhanced Fitness Coaching: From Exercise Form to Posture and Body Composition Analysis, Department of Computer Vision, Mohamed bin Zayed University, 2024.