
Furthermore, the related search term approach
supports customers in exploring the product cata-
log more comprehensively. By exposing users to
a broader range of relevant queries, the system en-
courages the discovery of products that may not have
been initially considered. This can lead to increased
engagement and potentially higher conversion rates,
benefiting both customers and the e-commerce plat-
form. The method, therefore, plays a crucial role in
optimizing the search functionality, ultimately driving
customer retention and loyalty.
2 RELATED WORKS
Research on related search terms also referred to
as query suggestion, query auto-completion (QAC),
query reformulation, or query expansion spans clas-
sical IR methods, graph- and log-mining approaches,
neural sequence models, and large-scale production
systems. Below, we synthesize the literature most rel-
evant to building and evaluating a system that surfaces
related search terms.
2.1 Query Suggestion from Logs and
Graphs
Early approaches exploit large-scale query and click
logs to model transitions between queries and to mine
semantically related alternatives. The query-flow
graph (QFG) represents queries as nodes and session-
based transitions as edges; random walks or edge
weights yield suggestions, (Boldi et al., 2009), (Boldi
et al., 2008), (Bai et al., 2011). A complementary line
uses query–document click bipartite graphs with hit-
ting time to balance semantic similarity and tail cov-
erage, (Mei et al., 2008). These methods established
that sequence context and click feedback enable high-
quality related suggestions beyond co-occurrence sig-
nals.
Auto-complete and prefix-sensitive variants ex-
tend suggestion to the character-level prefix setting,
where candidate generation and ranking must be real-
time. Classical works show that adding recent-query
context substantially improves short-prefix predic-
tions, (Bar-Yossef and Kraus, 2011). Surveys summa-
rize heuristic and learning-to-rank families, temporal
drift handling, and personalization, (Cai and de Rijke,
2016).
2.2 Context-Aware and Neural
Sequence Models
Generative neural models capture multi-query session
context to suggest the next reformulation. The hi-
erarchical recurrent encoder–decoder (HRED) con-
ditions on the sequence of prior queries, outper-
forming pairwise methods on next-query prediction,
(Sordoni et al., 2015). Subsequent work explores
deep language models for low-latency QAC, (Wang
et al., 2018) and integrates temporal/user features in
learning-to-rank frameworks (e.g., Hawkes/Markov
processes and user models), (Li et al., 2017), (Li et al.,
2015), (Kharitonov et al., 2013), (Cai et al., 2016).
2.3 Query Expansion and
Reformulation
To mitigate vocabulary mismatch, query expansion
augments the original query using pseudo-relevance
feedback (PRF) and relevance models. Relevance-
based language models (RM1/RM3) and PRF remain
strong baselines, (Lavrenko and Croft, 2001), (Abdul-
Jaleel et al., 2004; Carpineto and Romano, 2012; Za-
mani and Croft, 2011). Neural document/query ex-
pansion further improves first-stage recall: doc2query
and docTTTTTquery expand documents with pre-
dicted queries using sequence-to-sequence models,
boosting downstream ranking (Nogueira et al., 2019;
Nogueira and Lin, 2019). Recent studies revisit PRF
with modern embeddings and classification signals
(Lin et al., 2019; Wang et al., 2022). While expan-
sion targets retrieval effectiveness rather than UI sug-
gestions, the same candidates are valuable as related
search or people also search for terms.
2.4 Evaluation and Objectives
Offline evaluation commonly relies on histori-
cal logs with held-out user actions, measuring
top-k acceptance or reformulation success; user-
model–based metrics can better reflect utility in
QAC/QS, (Kharitonov et al., 2013; Cai and de Rijke,
2016). For related-term widgets (”People also search
for”), diversity and intent coverage are important; di-
versification for QAC has been studied explicitly, (Cai
et al., 2016). Counterfactual learning-to-rank for sug-
gestions from implicit feedback is increasingly rel-
evant to mitigate presentation bias (see surveys and
bandit LTR literature referenced therein).
ICEEECS 2025 - International Conference on Advances in Electrical, Electronics, Energy, and Computer Sciences
32