
2 RELATED WORKS
Product ranking in e-commerce has evolved sig-
nificantly, transitioning from rule-based heuristics
and standalone retrieval systems to learning-to-rank
(LTR) models that integrate heterogeneous signals.
Early approaches, such as collaborative filtering and
content-based ranking, often decoupled retrieval and
ranking stages. Modern architectures now favor uni-
fied pipelines that jointly optimize both tasks to align
with user preferences and business objectives (Kabir
et al., 2024).
LTR models, especially those based on gradient
boosting, have gained prominence for their predic-
tive strength and flexibility in handling mixed fea-
ture types. XGBoost (Chen and Guestrin, 2016) and
LightGBM (Ke et al., 2017) are frequently employed
in large-scale ranking tasks due to their scalability and
regularization techniques. CatBoost (Prokhorenkova
et al., 2018), in particular, excels at handling categor-
ical variables without preprocessing, making it well-
suited for e-commerce data with diverse categorical
attributes. In parallel, ensemble methods like Ran-
dom Forests and Extra Trees (Geurts et al., 2006)
serve as strong baselines for both model interpretabil-
ity and feature importance estimation. HistGradient-
Boosting, available via scikit-learn, offers computa-
tional efficiency by combining histogram-based train-
ing with support for monotonic constraints and miss-
ing values.
Despite the advancements in tree-based models,
logistic regression continues to be widely used in
real-time production environments for its low infer-
ence latency, simplicity, and well-calibrated proba-
bilistic outputs. When paired with systematic fea-
ture selection techniques—such as sequential forward
selection and Weight of Evidence (WoE) (Raymaek-
ers et al., 2021) binning—logistic regression achieves
strong performance while maintaining interpretability
(Loukili et al., 2023).
Beyond product recommendation, similar ranking
strategies have been applied to other personalization
tasks. For instance, a recent study on homepage ban-
ner optimization demonstrates that click prediction-
based ranking using logistic regression leads to mea-
surable improvements in click-through and conver-
sion rates (Keskin et al., 2024a). This application fur-
ther supports the viability of interpretable models in
latency-sensitive production systems.
Finally, fairness and transparency concerns are
increasingly relevant in commercial ranking. Stud-
ies have revealed that certain platforms may intro-
duce systemic biases—such as favoring private-label
or sponsored items—through opaque ranking poli-
cies. These findings underscore the importance of
explainability and bias-aware evaluation in deployed
recommendation models.
3 DATA COLLECTION
To support ”Frequently Bought Together” (FBT) rec-
ommendations, we construct a training dataset by in-
tegrating user interaction history, candidate recom-
mendations, and product metadata. As illustrated in
Figure 1, users generate both recommendation ex-
posures and order events, which are then merged
with product-level features to form labeled product
pairs. Candidate products are first retrieved using
embedding-based similarity, ranked by (1−distance).
Positive labels are assigned when the candidate was
co-purchased with the main product in the same order,
while negatives are drawn from unpurchased but rec-
ommended items. Data is split chronologically into
70% training, 15% validation, and 15% test to simu-
late real-world deployment.
Figure 1: Overview of the data collection and merging
pipeline.
The dataset integrates features from multiple
sources, including product catalog (category hierar-
chy, brand, merchant, reviews), pricing (listing prices,
view events, and historical orders), and user engage-
ment metrics (views, clicks, sales across six plat-
form touchpoints). Performance signals such as click-
through rate (CTR), conversion rate (CR), and per-
centile rankings are also incorporated.
Table 1 summarizes the groups. Prices follow a hi-
erarchical imputation strategy listing → view → order
→ default and are log-transformed to reduce skew.
Table 1: Overview of feature groups used in the ranking
model.
Feature Group # Features Source
Similarity & Position 2 Embedding, rank
Categorical Matching 5 Category, brand
Engagement Metrics 12 Views, clicks
Performance Ratios 6 CTR, CR
Commercial Signals 8 Price, reviews
Derived Features 4 Composites
ICEEECS 2025 - International Conference on Advances in Electrical, Electronics, Energy, and Computer Sciences
244