GLOSSARY

Learning to Rank (LTR)

Learning to Rank trains a model to order results using many features, not just text match. In stores, LTR blends relevance with signals like stock, rating, margin, and behavior to lift CTR and sales.

What is Learning to Rank?

Learning to Rank (LTR) uses machine learning to combine features into a scoring function that orders results. Approaches include pointwise (regression), pairwise (e.g., LambdaMART, BPR), and listwise objectives (e.g., ListNet, SoftNDCG).

How It Works (quick)

  • Features: Text scores (BM25/phrase), semantic signals (vectors), price, margin, rating, popularity, freshness, availability, personalization features.
  • Labels: Clicks/add-to-carts/purchases with position-bias corrections (IPS, randomized swaps, interleaving) and golden sets.
  • Training: Split by time/category; avoid leakage; tune on NDCG/MRR.
  • Serving: Retrieve candidates → compute features → model score → apply caps/diversity → render.
  • Monitoring: Track CTR@k, conv, margin per session, drift, and tail latency.

Why It Matters in E-commerce

  • Business-aware relevance: Balances textual match with quality and profitability.
  • Adaptivity: Learns from behavior while respecting guardrails.
  • Granularity: Per-category/locale models capture differences (fashion vs electronics).

Best Practices

  • Simple → complex: Start with LambdaMART/XGBoost; add deep re-rankers if latency allows.
  • Feature store: Version features with lineage; keep a human-readable baseline for rollback.
  • Bias handling: Use randomization/interleaving; apply IPS or counterfactual learning.
  • Constraints: Enforce in-stock, brand caps, and compliance before scoring.
  • Safety KPIs: Monitor returns, OOS clicks, and assortment coverage, not just CTR.
  • A/B discipline: Gate launches with golden set + online tests.

Challenges

  • Data leakage & drift: Future data or editorial boosts sneaking into training; seasonality shifts.
  • Cold start: Sparse feedback for new SKUs; blend with content-based scores.
  • Explainability: Complex models are hard to debug; keep feature attributions.

Examples

  • Search “merino base layer”: LTR elevates items with high rating/low return rate, in the user’s size, still respecting text match.
  • Category “sneakers”: Per-size availability and recent CTR carry more weight than in outerwear.

Summary

LTR learns a ranking that respects relevance and business quality. Start with interpretable models, handle bias, enforce constraints, and measure both engagement and profitability to ship gains safely.

FAQ

Do I still need BM25?

Yes—use it for recall and as a key feature; LTR reorders candidates.

When to personalize?

Include opt-in features (size/brand affinity) with privacy guardrails.

How often to retrain?

Weekly for behavior shifts; ad-hoc after big taxonomy or season changes.