There was a moment early in my career when a search result surprised me in the wrong way. 

I had typed a straightforward query. The results were relevant. They were even personalized in ways the system could justify. But my first reaction wasn’t delight. It was discomfort. I hadn’t told the product that much about myself, and yet it seemed to know more than I expected. 

The team was proud of the outcome. By internal measures, it was working. I remember sitting in the review, listening to discussions about lift and confidence scores, while struggling to explain a reaction that was not captured in any dashboard. The system was correct, but the experience felt wrong

That moment stayed with me because it captured a tension product teams rarely name openly. Personalization can improve relevance while quietly eroding trust. 

Search is where this tension becomes hardest to ignore. 

Users approach search with a specific mental model. They believe search reflects their intent in the moment. Not their identity. Not a long trail of historical behavior. Not a probabilistic profile assembled over time. 

That expectation makes search fundamentally different from feeds, recommendations, or ads. In those surfaces, users expect persuasion. In search, they expect responsiveness. Even when neutrality is imperfect, the illusion matters. 

I once worked on a product where we reused personalization logic that had performed well in a recommendation surface and applied it directly to search. From a systems perspective, the move made sense. The signals were strong. The model confidence was high. Engagement initially improved. 

But qualitative feedback shifted almost immediately. Users described results as “oddly specific” and “too confident for what I typed.” The logic had not changed. The surface had. 

Search amplifies personalization decisions because users experience them as answers, not suggestions. When those answers feel assumptive, trust erodes quickly. 

Where teams accidentally cross the line 

Most personalization failures don’t come from bad intent. They come from reasonable technical decisions made without enough product restraint. 

I’ve seen teams personalize too early, before a user has provided enough context within a session. I’ve seen systems overweight historical signals even when real-time intent was ambiguous. And I’ve seen personalization rules compound quietly, each one defensible on its own, until the experience became narrow and rigid. 

In one project, we noticed strong engagement from users who had previously shown interest in a specific category. The ranking system began to favor that signal more aggressively. Over time, search results clustered tightly around it. 

On paper, relevance improved. But a pattern emerged in support tickets and user research. People felt the product had “made up its mind” about them. One user wrote that searching felt less like exploration and more like confirmation. 

Nothing was technically broken. The models were behaving as designed. The failure was that we had allowed inferred preference to override expressed intent too quickly. 

That is a product decision, not a modeling error. 

Exploring AI bias and how to design products with intention
Uncover the biases AI surfaces, and how these can reflect our blind spots and guide us to better product design solutions.

The difference between helpful and uncomfortable 

After years of working on search products, I’ve found it useful to think about personalization through a simple lens. 

Helpful personalization tends to feel: 

  • Closely tied to what the user just did 
  • Grounded in observable session behavior 
  • Reversible over time 

Uncomfortable personalization tends to feel: 

  • Based on assumptions the user never explicitly offered 
  • Difficult to explain without referencing internal systems 
  • Persistent in ways the user cannot influence 

This isn’t a framework. It’s a gut check. 

In ranking reviews, I started asking a simple question: If a user paused on this result and asked why it appeared, could we explain it in plain language? When the answer relied on phrases like “the model inferred” or “similar users tended to,” it was usually a sign we were crossing a line. 

The moment personalization requires a long internal explanation, trust is already under pressure. 

AI ethics and governance framework
Your go-to framework for ethical and transparent AI practices.

Where AI and ML raise the stakes 

Modern search personalization is powered by increasingly capable machine learning systems. These models can identify subtle correlations across millions of interactions and adapt continuously as new data arrives. 

That capability changes the product equation. 

I’ve been in rooms where a model’s confidence score climbed while the team’s confidence in the experience declined. The system was certain. The product intuition was not. In those moments, it became clear that better inference does not automatically lead to better judgment. 

Models optimize toward the signals we give them. If those signals overrepresent historical behavior, the system will amplify history. If they overweight engagement, the system will narrow outcomes. The model isn’t wrong – it’s obedient. 

As AI systems become more powerful, the cost of getting these decisions wrong increases. Personalization errors scale faster. They become harder to notice. And they are often justified by metrics that look healthy in isolation. 

The more capable the system, the more intentional the product constraints need to be. 

Unpacking the craft of an applied machine learning PM
Bridge the gap between ML models and user experience. Learn the craft of an applied ML product manager: turning research into real-world impact.

Guardrails that actually help 

The most effective personalization guardrails I’ve seen are surprisingly simple, but they require discipline. 

One is progression. Let personalization earn its way into the experience instead of asserting itself immediately. Early search results should privilege clarity and diversity. Deeper personalization can follow once intent becomes clearer. 

Another is decay. Historical signals should fade unless they are repeatedly reinforced. Interests change. Systems should reflect that fluidity instead of freezing users in place. 

A third is escape. Users don’t need fine-grained controls for every signal, but they do need to feel that a new query resets assumptions. Search should feel responsive, not judgmental. 

In one case, we deliberately weakened a personalization rule that had strong short-term performance. The decision was uncomfortable, and engagement dipped slightly at first. But over time, exploration increased, complaints declined, and user trust indicators improved. 

That tradeoff was invisible in any single experiment. It only became clear when we stepped back and evaluated the system as a whole. 

Beyond the AI hype: Asking the right questions to drive meaningful product innovation
Is AI the right move for your product? Learn how to ask the right questions when implementing AI into your products to drive meaningful innovation.

Why this is a product responsibility 

Personalization systems learn from data. Engineers build pipelines. Models adapt continuously. But product managers decide where personalization applies, how visible it is, and when it stops. 

Those decisions shape how much power a product quietly holds over its users. In search, that power is amplified because visibility feels neutral even when it’s not. 

Treating personalization as a feature undersells its impact. It’s closer to a form of influence. Influence without trust doesn’t scale. The best search experiences don’t make users feel dramatically understood. They make them feel respected. That difference is subtle, but it determines whether users lean in or pull away. 

Personalization will continue to improve technically. The harder work is ensuring it continues to earn the right to exist at all.