Pre-trained language models (PLMs) are widely used to derive semantic representations from item metadata in recommendation and search. In sequential recommendation, PLMs enhance ID-based embeddings through textual metadata, while in product search, they align item characteristics with user intent. Recent studies suggest task and domain-specific fine-tuning are needed to improve representational power. This paper challenges this assumption for e-commerce applications, showing that Generalist Text Embedding Models (GTEs), pre-trained on large-scale corpora, can guarantee strong zero-shot performance without specialized adaptation. Our experiments on popular e-commerce benchmarks demonstrate that GTEs outperform traditional and fine-tuned models in both sequential recommendation and product search. We attribute this to a superior representational power, as they distribute features more evenly across the embedding space. Finally, we show that compressing embedding dimensions by focusing on the most informative directions (e.g., via PCA) effectively reduces noise and improves the performance of specialized models. To ensure reproducibility, we provide our repository at https://github.com/sisinflab/GTE-Zero-Shot-Recsys.

Do We Really Need Specialization? Evaluating Generalist Text Embeddings for Zero-Shot Recommendation and Search / Attimonelli, Matteo; De Bellis, Alessandro; Pomo, Claudio; Jannach, Dietmar; Di Sciascio, Eugenio; Di Noia, Tommaso. - (2025), pp. 575-580. ( 19th ACM Conference on Recommender Systems, RecSys 2025 Prague, Czech Republic ) [10.1145/3705328.3748040].

Do We Really Need Specialization? Evaluating Generalist Text Embeddings for Zero-Shot Recommendation and Search

Attimonelli, Matteo
Primo
;
2025

Abstract

Pre-trained language models (PLMs) are widely used to derive semantic representations from item metadata in recommendation and search. In sequential recommendation, PLMs enhance ID-based embeddings through textual metadata, while in product search, they align item characteristics with user intent. Recent studies suggest task and domain-specific fine-tuning are needed to improve representational power. This paper challenges this assumption for e-commerce applications, showing that Generalist Text Embedding Models (GTEs), pre-trained on large-scale corpora, can guarantee strong zero-shot performance without specialized adaptation. Our experiments on popular e-commerce benchmarks demonstrate that GTEs outperform traditional and fine-tuned models in both sequential recommendation and product search. We attribute this to a superior representational power, as they distribute features more evenly across the embedding space. Finally, we show that compressing embedding dimensions by focusing on the most informative directions (e.g., via PCA) effectively reduces noise and improves the performance of specialized models. To ensure reproducibility, we provide our repository at https://github.com/sisinflab/GTE-Zero-Shot-Recsys.
2025
19th ACM Conference on Recommender Systems, RecSys 2025
Generalist Text Embedding Models; Product Search; Sequential Recommendation
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Do We Really Need Specialization? Evaluating Generalist Text Embeddings for Zero-Shot Recommendation and Search / Attimonelli, Matteo; De Bellis, Alessandro; Pomo, Claudio; Jannach, Dietmar; Di Sciascio, Eugenio; Di Noia, Tommaso. - (2025), pp. 575-580. ( 19th ACM Conference on Recommender Systems, RecSys 2025 Prague, Czech Republic ) [10.1145/3705328.3748040].
File allegati a questo prodotto
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1753388
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact