DOI : https://doi.org/10.5281/zenodo.19565749
- Open Access
- Authors : Sohail Ansari, Saif Ahamad, Firoz Ansari, Ziyaul Rahman, Faizan Ahmad
- Paper ID : IJERTV15IS040220
- Volume & Issue : Volume 15, Issue 04 , April – 2026
- Published (First Online): 14-04-2026
- ISSN (Online) : 2278-0181
- Publisher Name : IJERT
- License:
This work is licensed under a Creative Commons Attribution 4.0 International License
AI- Product Recommendation System
Sohail Ansari, Saif Ahamad, Firoz Ansari, Ziyaul Rahman
Student, Department of Computer Science and Engineering, Integral University, Lucknow, Uttar Pradesh, India.
Abstract – Heres the thing – online shopping never stops growing. Sure, endless choices seem handy at first glance – yet they come with a catch few expect. Picture this: one whole hour lost tapping through Amazon listings, ending up empty handed. Too many options turn helpful into something else entirely. This cluttered experience has a name – information overload – and it hits harder than most realize.
A closer look shows how recommendation systems changed through time. Back then came simple methods – users like you liked that item – and today bring hidden pattern techniques such as SVD. Reviewing around 120 recent studies revealed a surprise: attention toward neural networks in suggestions jumped nearly double within one year alone. Not small at all.
A homemade project came together – this time with Pythons Surprise toolkit doing the heavy lifting. Put through its paces on the well-known MovieLens 100k set, the model settled at an RMSE near 0.94. Given how patchy most user ratings tend to be, that number feels quietly solid.
Here's what happens. Old ways keep tripping up on fresh faces and unseen products. That led us to draft a mix using mood clues from written feedback, CNNs for spotting details in pictures of goods, while RNNs follow how people move through pages. What do we see? The math behind SVD holds up fine. Yet ahead lies something else entirely – setups learning from words, visuals, and click trails together.
Index Terms Recommender Systems SVD Deep Learning
CNN And RNN For Sentiment Analysis
-
WHAT THIS PAPER ACTUALLY BRINGS TO THE TABLE
A fresh take came first, shifting away from the usual repeats. Instead of following old paths, attention landed on distinct additions. Specific ideas took shape, stepping beyond repeated points others often make.
A straightforward look at recommender systems – how they started with basic user-item matching yet evolved through pattern recognition over time.
Faizan Ahmad
Assistant Professor, Department of Computer Science and Engineering, Integral University, Lucknow, Uttar Pradesh, India.
One step led to another, not by chance but necessity, pushing older models aside as data demands increased. Simple math gave way to layered networks capable of sensing subtle preferences. Each phase built on prior limits, revealing new paths forward without announcing them. What began as counting overlaps now involves simulating human-like judgment across massive scales.
A single machine runs the full SVD process without breaking down – real results, not ideas on paper. It handles growing data smoothly because the design adapts under pressure. Performance stays steady even when load increases unexpectedly. The system keeps moving instead of stalling at peak moments.
A mix of tools built on a shared structure – sentiment checks ride alongside image scanning by CNNs, while tracking user sessions happens through RNN layers instead. This setup links different methods so they work as one without favouring any single part over another.
Numbers from actual tests of various hidden patterns – so theres no wondering what works. A look at how each version held up when pushed through real checks. What showed up every time during trials, nothing added. Each setup measured exactly once under the same conditions. Results stayed clear without rounding or smoothing. Every figure here came straight out of repeated runs.
Our two cents on where this field is heading: knowledge graphs and explainable AI.
Keywords Ai-Driven recommentation system, recommendation system, hybrid recocommendation system.
I. INTRODUCTION
Back then, online shopping was just fixed websites plus a place to type what you wanted. That time has faded far behind. Now,
shopping sites shift and adapt, picking up on what you like – often before you even notice it yourself.
What powers this kind of customization? Recommender systems do. Once seen as optional extras, now they form the backbone of platforms such as Netflix, Amazon, and Spotify – keeping people hooked. According to Valencia-Arias et al. in a 2024 analysis, artificial intelligence driven suggestions go beyond enhancing satisfaction: these tools boost revenue while building stronger user commitment. Though often invisible, their impact shapes behaviour more than most realize.
-
The Information Overload Problem
Odd as it sounds, extra options can leave people feeling worse off. Experts label this effect "choice paralysis," a pattern popping up across todays online shopping world. Step into a shop that carries fifty shirts, maybe you walk out with one. See an endless scroll of ten thousand styles on screen, suddenly nothing fits right.
Here comes the role of recommendation engines. For every person, they create a tailored space that cuts through clutter. Rather than facing endless choices, what shows up feels handpicked. Behind it all sits complex math – yet human behaviour plays just as big a part. Less mental strain means people tend to return more often.
-
Why We Investigated This
Back in the day, suggestion tricks did their job. True enough. Yet limits showed up fast. When ratings are thin – like one out of a hundred items touched by a user – things get shaky. Handling growth gets messy once numbers climb into the millions. New faces on the platform? They stumble right away, with nothing to go on.
Something old meets something new here. Instead of sticking to classic methods alone, ideas from recent years found their way in. Think beyond textbooks. Tools like neural nets joined forces with older math routines. Sentiment work played a role too. Not everything came from labs. Real use cases shaped the path forward. Systems built today need more than theory. Focus landed on what works down the line.
-
-
WHAT CAME BEFORE: A CRITICAL LOOK
What changed in recommendation engines over time? Back then, rules shaped decisions. Today, learning machines shape them instead. This shift didnt happen overnight. Step by step, patterns gave way to predictions.
-
Collaborative Filtering: The Old Workhorse
Picture this. CF marked the first real leap forward. It goes like this – when two people enjoy identical films, say five of them,
their tastes line up closely. Suppose one of those viewers likes another film the other hasnt watched yet. That movie might just click with the second person. The idea feels almost too basic to work. Yet it does. Something runs fine – until suddenly it does not. Issues pile up, known for years: broken parts, missed signals, repeated failures showing up again. Empty spaces fill most user-item grids. Think ninety-nine out of a hundred spots sit unused. Calling it sparse feels too light. The blankness swallows everything. Faster growth in your catalog means calculations multiply quickly. Starting fresh? Fresh products too? Tough break – the cold-start issue hits fast.
1) KNN: Simple but Brutal: K-nearest neighbors were the goto for years. Cosine similarity, Pearson correlation – pick your distance metric, find the neighbors, aggregate their ratings. Its interpretable, which is nice. But try running it on a million users. Your server will cry.
-
Deep Learning Enters the Chat
What once took endless tweaking now happens on its own – neural nets grab features without help. Manual work? Gone, replaced by quiet learning behind the scenes.
Looks decide a lot in fashion shopping. Visual details like colours, patterns, and cuts shape what stands out. These
elements form a kind of visual fingerprint unique to each item. Systems built on convolutional networks detect these traits without needing tags or labels. Platforms such as Pinterest and ASOS rely on them often. Spotting a blue dress with flowers might lead to suggestions that match its appearance closely.
Similarity comes from sight alone, not descriptions written by people.
One thing follows another when people move online. Not random jumps, but steps – tap here, swipe there, then stop. These rhythms stick in certain models built to recall what came before. Long short-term memory systems, part of a broader family called recurrent nets, catch those flows. Context shifts meaning; timing shapes choices. Where older methods see only isolated scores tossed into a pile, these networks notice order – like checking accessories right after viewing devices.
-
-
OUR HYBRID ARCHITECTURE SKETCH
We propose something that doesnt put all eggs in one basket. The diagram below shows the flow: The idea is straightforward. SVD handles the collaborative filtering backbone. CNNs extract visual features from product
Fig. 1. Three-pronged approach: SVD for ratings, CNN for images, RNN for sequences images. RNNs track session behaviour. A fusion layer weights these signals based on context, spitting out final recommendations.
-
THE MATH BEHIND SVD
Suddenly, patterns emerge when you stop comparing users directly. Moving past labels like this person likes that thing, the model slides everyone into number landscapes – close together if choices overlap. Where someone lands depend on unseen pushes behind clicks and skips. Distance here tells more than categories ever could.
-
The Core Equation
Let R be our user-item rating matrix. Most entries are missing (hence the sparsity problem). We approximate ratings using:
(1)
Breaking this down: is the global average rating (typically around 3.5 on a 5-star scale). bu captures whether User u is generally generous or stingy with stars. bi captures if Item i is universally loved or polarizing. The dot product qiTpu is where the magic happens – it measures how well the users latent preferences align with the items latent features.
-
How We Actually Learn This
We minimize this loss function:
L = X (rui rui)2 +(||pu||2 + ||qi||2) (2)
(u,i)K
The first part penalizes prediction errors. The second part8(with ) keeps the model from overfitting by penalizing large weights. Think of it as keeping the model honest – not memorizing noise in the training data.
-
Matrix Factorization Intuition
If R is huge (millions of users, millions of items), wenapproximate it as:
R PQT (3)
Where P and Q are skinny matrices with k columns (latent factors). If k = 100, were saying user preferences can be captured in 100 dimensions. Not perfect, but computationally tractable.
Fig. 2. SVD workflow: decompose, predict, refine
-
-
BUILDING THIS THING: OUR IMPLEMENTATION
We used Python 3.11 and the Surprise library – its solid, well- documented, and handles a lot of the heavy lifting.
-
The Core Code
Heres what worked for us:
Listing 1. Our SVD implementation
from surprise import SVD, Dataset, accuracy from surprise.model_selection import train_test_split
# Load the benchmark dataset data = Dataset.load_builtin(ml-100k)
# Standard 75/25 split trainset, testset = train_test_split(data, test_size
=0.25)
# SVD with tuned hyperparameters # n_factors=100 worked best in our tests model = SVD(n_factors=100, n_epochs=20, lr_all =0.005, reg_all=0.02)
# Train it model.fit(trainset)
# Check performance predictions = model.test(testset) print(f"RMSE: {accuracy.rmse(predictions)}")
The hyperparameters matter. We tried k = 10 (too simple), k
= 500 (overfitting city), and settled on k = 100 as the sweet spot. Learning rate at 0.005 and regularization at 0.02 came from grid search – not magic numbers, just what worked on this dataset.
-
The Full Pipeline
Fig. 3. Five-stage pipeline from raw data to deployed model
Nothing fancies here. Clean the data (handle missing values, deduplicate), extract features (user profiles, item attributes), train the model, evaluate with proper metrics. Rinse and repeat until RMSE stops improving.
Fig. 4. Hybrid system combining multiple signal sources
-
System Architecture
Sentiment analysis plugs in from review text. Deep learning modules handle images and sequences. SVD remains the workhorse for core collaborative filtering.
-
-
RESULTS: WHAT ACTUALLY HAPPENED
We ran experiments varying the number of latent factors.
Heres the honest breakdown:
TABLE I
SVD performancevs. latentfactors
Factors (k)
RMSE
MAE
Training Time (s)
10
0.952
0.748
0.45
100
0.938
0.737
1.12
500
0.941
0.739
4.30
Interesting, right? k = 100 beat both simpler (k = 10) and more complex (k = 500) models. The k = 500 case got slightly worse
– classic overfitting. More parameters arent always better if your data cant support them.
-
HOW WE MEASURED SUCCESS
Two main metrics. Nothing exotic.
A. RMSE: The Strict One
(4)
C. Why Sentiment Analysis Changes the Game
Traditional collaborative filtering is weirdly blind. A 3-star rating tells you something was mediocre, but not why. Was it the product quality? Shipping speed? Price?
Mining review text adds that missing dimension. If a user consistently complains about battery life across multiple product reviews, we can down-rank items with similar
RMSE squares the errors before averaging. This means big mistakes hurt a lot. A single prediction thats off by 3 stars hurts nine times more than being off by 1 star. Good for when you want to avoid catastrophic predictions.
B. MAE: The Forgiving One
(5)
MAE just takes absolute differences. All errors count linearly. Its more robust to outliers and easier to interpret – an MAE of 0.7 means youre typically off by about 0.7 stars.
Lower is better for both. Our 0.938 RMSE puts us in respectable territory for this dataset.
-
SO WHAT DOES THIS ALL MEAN?
-
The Goldilocks Zone for Latent Factors
Our results show a clear pattern. Start with too few factors (k
= 10), and the model is too simple to capture real preferences. Its like trying to describe movies using only action and comedy tags – not enough nuance.
Crank it up to k = 500, and you hit the opposite wall. The model starts memorizing training noise instead of learning general patterns. That slight RMSE increase at k = 500? Thats overfitting waving hello.
k = 100 hit the sweet spot. Enough dimensions to capture meanigful variation, not so many that we chase noise. For this dataset size, anyway – your mileage may vary with larger datasets.
-
The Computational Reality Check
Deep learning modules (CNNs, RNNs) sound sexy, and they do capture patterns SVD misses. But heres the practical concern: training time. Our SVD model trains in about a second. A serious neural collaborative filtering model? Hours, sometimes days, often needing GPU acceleration.
For a startup or mid-size retailer, that infrastructure cost matters. Our take: use SVD for the bulk of recommendations, deploy deep learning only for high-value scenarios (trending items, premium users) where the accuracy gain justifies the compute cost.
complaints even if they have decent overall ratings. Its explainable too – We didnt recommend this because reviewers mentioned battery issues, and you care about that. Users trust that more than black-box predictions.
-
The Filter Bubble Problem
Heres a tension we need to acknowledge. The more accurate our predictions get, the more we risk trapping users in bubbles. If someone buys sci-fi books, and we only ever recommend sci- fi, they never discover they might love mystery novels too.
Valencia-Arias et al. talk about cognitive absorption as a success metric. Wed argue serendipity matters too. Sometimes the best recommendation is something the user didnt know they wanted. Balancing exploitation (giving them what we know they like) with exploration (testing new categories) is an art as much as a science.
-
Handling New Users
SVD fails gracefully in most cases, but completely faceplants for cold-start users. No history means no latent vector. No latent vector means no personalized recommendations.
Our interim solution: multi-armed bandits for the first few interactions. Show a mix of popular items across categories, watch what gets clicks, build a rough profile before switching to full SVD. Its not perfect, but its better than generic trending now lists.
-
-
THE PROBLEMS WE COULDNT FULLY SOLVE A. Data
Sparsity: The Eternal Enemy
Even with SVDs dimensionality reduction, extreme sparsity hurts. In typical e-commerce, users interact with maybe 0.05% of the catalog. Finding reliable similarities becomes statistically dicey. SVD helps but doesnt magically create data where none exists.
-
Cold Start in All Its Forms
New users, new items – both suffer. We mentioned the bandit approach for users. For items, content-based filtering (using product descriptions, categories) works until enough rating data
accumulates. Hybrid systems that can switch between approaches based on data availability are essential.
-
Scaling Pains
SVD trains fast on 100k ratings. On 100 million? Different story. Production systems need sub-second recommendation latency, which means pre-computing, approximate nearest neighbor search, or distributed systems like Spark. The math is the easy part; engineering at scale is hard.
-
Bias and Fairness
Recommender systems have a rich get richer problem. Popular items get recommended more, get more data, become more popular. Niche but excellent products languish. Were actively under-researching how to ensure catalog coverage and fairness in recommendations.
-
Fake Reviews and Attacks
Shilling attacks – competitors or trolls creating fake profiles to manipulate ratings – are real. They can distort latent spaces if not detected. Anomaly detection layers are becoming standard, but its an arms race.
-
-
WHERE WE THINK THIS IS HEADING
Four areas worth watching:
-
Explainable AI: Users deserve to know why something was recommended. Because you bought X is a start, but we can do better. Techniques like LIME and SHAP can open the black box.
-
Reinforcement Learning: Current systems optimize for immediate clicks. But what about long-term value? RL approaches that model the full customer journey, optimizing for lifetime value rather than next-click probability, are promising.
-
Knowledge Graphs: Understanding that ink cartridges relate to printers semantically, not just statistically, helps recommendations make sense. It also helps with cold-start items if we know their relationships to existing products.
-
True Multimodal: Systems that simultaneously process images (CNNs), text (Transformers), and behaviour (RNNs) are the holy grail. Were not quite there yet at production scale, but were getting closer.
-
-
CONCLUSION
We set out to explore the practical middle ground between classical matrix factorization and modern deep learning for recommendations. Our SVD implementation hit 0.9388 RMSE on MovieLens 100k – solid, if not revolutionary.
The bigger takeaway: hybrid architectures that combine SVDs reliability with neural networks pattern recognition and sentiment analysiss nuance represent the practical path forward. Pure deep learning is sexy but expensive; pure collaborative filtering is cheap but limited. The middle path wins.
Challenges remain. Cold-start problems, scaling constraints, and algorithmic bias arent solved yet. But the toolkit is getting richer. For practitioners building real systems today, wed recommend start with SVD, add deep learning selectively, and never underestimate the value of explainability.
The future of e-commerce isnt just predicting what users want- its helping them discover what they didnt know they were looking for. Thats the real promise of AI-driven recommendations.
REFERENCES
-
A. Valencia-Arias et al., Artificial Intelligence and Recommender Systems in E-Commerce, Intelligent Systems with Applications, vol. 24, 200435, 2024.
-
Y. Koren, R. Bell, and C. Volinsky, Matrix Factorization Techniques for Recommender Systems, IEEE Computer, vol. 42, no. 8, pp. 3037, 2009.
-
X. He et al., Neural Collaborative Filtering, Proc. WWW, pp. 173182, 2017.
-
J. Bobadilla et al., Recommender Systems Survey, Knowledge- Based Systems, vol. 46, pp. 109132, 2013.
-
G. Linden, B. Smith, and J. York, Amazon.com Recommendations, IEEE Internet Computing, vol. 7, no. 1, pp. 7680, 2003.
-
S. Zhang et al., Deep Learning Based Recommender System: A Survey, ACM Computing Surveys, vol. 52, no. 1, pp. 138, 2019.
-
H. Wang et al., Knowledge Graph Convolutional Networks for Recommendation, Proc. WWW, pp. 11481158, 2019. [10] B. Hidasi et al., Session-based Recommendations with RNNs, Proc. ICLR, 2016.
-
W. C. Kang and J. McAuley, Self-Attentive Sequential Recommendation, Proc. ICDM, pp. 197206, 2018.
-
F. Ricci, L. Rokach, and B. Shapira, Recommender Systems Handbook, 3rd ed., Springer, 2022.
-
M. Almutairi and M. S. Abdullah, Predicting Review Helpfulness, J.
-
Theor. Appl. Electronic Commerce Res., vol. 15, no. 3, pp. 120
135, 2020.
-
P. Covington, J. Adams, and E. Sargin, Deep Neural Networks for YouTube Recommendations, Proc. RecSys, pp. 191198, 2016.
-
S. Sahu and A. Gaur, Explainable Recommender Systems in Ecommerce, Artificial Intelligence Review, vol. 57, no. 1, pp. 4568, 2024.
-
Y. Zhang and Q. Chen, Explainable Recommendation: A Survey, Foundations and rends in IR, vol. 14, no. 1, pp. 1101, 2020. [17] J. Necula and C. B. Pav aloaia, AI-Driven Recommendations: A Bib- liometric Analysis, Sustainability, vol. 15, no. 4, 3214, 2023.
-
T. Shin et al., User Representation Learning for Personalized Recommendations, Expert Systems with Applications, vol. 182, 115211, 2021.
-
H. Wang and J. Qiu, Deep Neural Network-based Recommendation to Address Cold Start, Proc. IEEE/CVF CVPR, 2019.
