Home Artificial Intelligence Beyond Accuracy: Embracing Serendipity and Novelty in Recommendations for Long Term User Retention What makes a great advice? The Shift to Diversity Metrics What are these “beyond accuracy” objectives? Concluding Remarks References

Beyond Accuracy: Embracing Serendipity and Novelty in Recommendations for Long Term User Retention What makes a great advice? The Shift to Diversity Metrics What are these “beyond accuracy” objectives? Concluding Remarks References

1
Beyond Accuracy: Embracing Serendipity and Novelty in Recommendations for Long Term User Retention
What makes a great advice?
The Shift to Diversity Metrics
What are these “beyond accuracy” objectives?
Concluding Remarks
References

RECOMMENDER SYSTEMS

An examination of the aspects that contribute to a great advice and long-term user retention

Composition by the creator using DALL-E

A Bond of Trust Formed at a Coffee Shop

You’re sitting in a coffee shop, savoring your favorite coffee variation (a cappuccino, after all) and engrossed in conversation with a friend. Because the conversation flows, the subject shifts to the newest gripping TV series that you just each have been hooked on. The shared excitement creates a bond of trust, to the extent that your friend eagerly turns to you and asks, “What should I watch next? Do you have got any recommendations?”

At that moment, you turn out to be the curator of their entertainment experience. You’re feeling a way of responsibility to preserve their trust and supply suggestions which can be guaranteed to captivate them. Moreover, you’re excited at the chance to, perhaps, introduce them to a rather recent genre or storyline they hadn’t explored before.

Photo by Thibault Penin on Unsplash

First, you tap into your understanding of your friend’s tastes and interests. You recall their fondness for intricate plot twists and dark humor; moreover, you realize they enjoyed crime dramas like “Sherlock” and psychological thrillers like “Black Mirror.” Armed with this data, you navigate your mental library of TV shows.

To play it protected?

You’re tempted to suggest a listing of shows which can be almost similar, with slight variations, to the one you had just been raving over, which encompass each crime and thrill. You furthermore may take into consideration how others with similar tastes have enjoyed these shows to narrow your decisions. In spite of everything, they’re practically to enjoy this set; it’s the protected and simple alternative. Nonetheless, you concentrate on that and don’t particularly need to depend on the protected and simple alternative.

You recall a recent sci-fi series that ingeniously blends mystery, adventure, and supernatural intrigue. Even though it deviates from their typical genre, you’re feeling confident it’s going to provide a refreshing and charming change of narrative.

The Long Tail Problem, Feedback Loop & Filter Bubbles

Advice systems aim to duplicate this process on a bigger scale. By analyzing vast amounts of knowledge about individuals’ preferences, behaviors, and past experiences, these systems strive to generate personalized recommendations that encompass the complexity of human decision-making.

Nonetheless, traditionally, advice systems have focused primarily — if not, solely — on and counting on the recommendations which can be guaranteed to satisfy (a minimum of, within the short term).

A method they do that is by prioritizing popular or mainstream content. In consequence, this popular content receives more exposure and interactions (), making a feedback loop that reinforces its prominence. Unfortunately, this often leaves lesser-known or area of interest content struggling to realize visibility and reach the intended audience ().

Photo by Writer

The truth is, there was a number of literature in the previous few years that examine “fairness” in advice systems. For instance, in “Fairness in Music Recommender Systems: A Stakeholder-Centered Mini Review”, Karlijn Dinnissen and Christine Bauer explore the difficulty of fairness in music recommender systems; they analyze gender fairness and recognition bias from the angle of multiple stakeholders e.g. the impact of recognition bias on the representation of artists.

Within the article, “Fairness in Query: Do Music Advice Algorithms Value Diversity?”, Julie Knibbe shares:

As a former product director at a streaming platform, I often receive questions like “do streaming services select to advertise popular artists over indies and area of interest music?” Intuitively, people assume that on these big platforms “the wealthy get richer.”

Afterward within the article, Knibbe also echoes the sentiment of Dinnissen and Bauer:

“Within the context of music advice […] fairness is usually defined by way of or . Streaming services are also a two-sided marketplace, which suggests that

Each sources highlight the twin nature of fairness in recommender systems, underscoring the importance of considering “impartial and just treatment” for users and content creators.

What does the perfect consequence appear to be?

Naturally, there exists an inherent imbalance within the distribution of content. A part of what makes the human experience wealthy lies inside its network intricacy; some content resonates with a broader audience, while others forge connections inside area of interest groups, developing a way of richness and personalization. The target shouldn’t be to artificially promote less popular content for the sake of it, striving for a uniform distribution. Reasonably, our aim is to surface area of interest content to individuals who genuinely relate and may appreciate the content creator’s work, thereby minimizing missed opportunities for meaningful connections.

In 2020, the research team at Spotify released an article titled, “Algorithmic Effects on the Diversity of Consumption on Spotify.” Of their research, they examined the connection between listening diversity and user outcomes.

Photo by Fath on Unsplash

They aimed to reply the questions: “How does diversity relate to necessary user outcomes? Are users who listen diversely kind of satisfied than those that listen narrowly?”

The researchers discovered that “users with diverse listening are between points than those with […] and”

Moreover, in keeping with Julie Knibbe:

“TikTok’s advice algorithm was recently mentioned among the many top 10 […] by MIT technology review. What’s revolutionary of their approach isn’t the algorithm itself — it’s the metrics they’re optimizing for, weighing in additional on diversity than other aspects.”

Subsequently, there may be a connection between the attribute of discoverability inside a platform and user retention. In other words, when recommendations turn out to be predictable, users might seek alternative platforms that provide a greater sense of “freshness” in content, allowing them to flee the confines of filter bubbles.

So how can advice systems emulate the thoughtfulness and intuition that you just employed in curating the proper suggestion in your friend?

Well, within the article, “Diversity, Serendipity, Novelty, and Coverage: A Survey and Empirical Evaluation of Beyond-Accuracy Objectives in Recommender Systems”, authors Marius Kaminskas and Derek Bridge highlight:

“Research into recommender systems has traditionally focused on accuracy […] nevertheless, it has been recognized that other advice qualities — resembling whether the list of recommendations is diverse and whether it accommodates novel items — can have a major impact on the general quality of a recommender system. Consequently […] the main target of recommender systems research has shifted to incorporate a wider range of ‘beyond accuracy’ objectives”

Diversity

Sifting through the literature in an try to understand what ‘’ is in recommender systems was brutal, as each article presented its own unique definition. Diversity could be measured each at the person levelor at the worldwide level We’ll go over 3 ways to conceptualize diversity, within the context of giving show recommendations to a friend.

Prediction diversity refers back to the measure of how varied the recommendations are inside a given set. Once you suggest a set of shows to your friend, prediction diversity assesses the extent to which the recommendations differ from each other by way of genres, themes, or other relevant aspects.

The next prediction diversity indicates a wider range of options inside the beneficial set, offering your friend a more diverse and potentially enriching viewing experience.

A method that is measured is through the use of which is the typical pairwise dissimilarity among the many beneficial items. Given the beneficial item list, the ILD is defined as follows:

User diversity, within the context of providing show recommendations to a friend, examines the typical diversity of all of the recommendations you have got ever given to that specific friend. It considers the breadth and number of content suggested to them over time, capturing the range of genres, themes, or other relevant aspects covered.

You can even assess user diversity by analyzing the typical dissimilarity between the item embeddings inside each set of recommendations per friend.

However, global diversity looks beyond a selected friend and assesses the typical diversity of all of the recommendations you have got given to any friend.

Sometimes, that is known as a mirrored image of advice uniformity or the crowding of recommendations.

A few metrics that you would be able to use to research global diversity include the Gini index and entropy.

Theadapted from the sector of income inequality measurement, could be used to evaluate the fairness and balance of advice distributions in advice systems. A lower Gini index indicates a more equitable distribution, where recommendations are spread evenly, promoting greater diversity and exposure to a wider range of content. However, a better Gini index suggests a concentration of recommendations on a number of popular items, potentially limiting the visibility of area of interest content and reducing diversity within the recommendations.

is a measure of the quantity of data contained within the advice process. It quantifies the extent of uncertainty or randomness within the distribution of recommendations. Just like the Gini index, optimal entropy is attained when the advice distribution is uniform, meaning that every item has an equal probability of being beneficial. This means a balanced and diverse set of recommendations. Higher entropy suggests a more varied and unpredictable advice system, while lower entropy indicates a more concentrated and predictable set of recommendations.

Coverage

Coverage is defined because the portion/proportion of possible recommendations the algorithm can produce. In other words, how well the recommendations cover the catalog of obtainable items.

For instance, let’s consider a music streaming platform with an unlimited library of songs spanning various genres, artists, and many years. The coverage of the advice algorithm would indicate how effectively it will probably cover everything of this music catalog when suggesting songs to users.

: This metric treats an item beneficial once as similar to an item that was beneficial hundreds of times

Novelty is a metric used to gauge the extent of newness or originality in beneficial items. It encompasses two points: user-dependent and user-independent novelty. User-dependent novelty measures how different or unfamiliar the recommendations are to the user, indicating the presence of fresh and unexplored content. Nonetheless, it has turn out to be increasingly common to discuss with the novelty of an item in a user-independent way.

To estimate novelty, one common approach is to think about an item’s popularity, measured as Item Rarity. This approach inversely relates an item’s novelty to its popularity, recognizing that less popular items are sometimes perceived as more novel as a consequence of their deviation from mainstream or widely-known decisions. By integrating this attitude, novelty metrics provide insights into the extent of innovation and variety present within the beneficial items, contributing to a more enriching and exploratory advice experience.

Unexpectedness (Surprise)

Surprise in advice systems measures the extent of unexpectedness within the beneficial items based on a user’s historical interactions. One method to quantify surprise is by calculating the cosine similarity between the beneficial items and the user’s past interactions. The next similarity indicates less surprise, while a lower similarity indicates greater surprise within the recommendations.

Discoverability in advice systems could be understood because the user’s ability to simply come across and find the recommendations suggested by the model. It’s akin to how visible and accessible the recommendations are inside the user interface or platform.

It’s quantified using a decreasing rank discount function, which assigns higher importance to recommendations at the highest ranks of the advice list and steadily decreases their weight because the rank position goes down.

Serendipity in advice systems encompasses two key points: unexpectedness and relevance.

Serendipity refers back to the occurrence of nice surprises or the invention of interesting and unexpected recommendations. To quantify serendipity, it’s calculated on a per-user and per-item basis using the formula:

By multiplying unexpectedness and relevance, the serendipity metric combines the weather of nice surprise and suitability. It quantifies the degree to which a advice is each unexpected and relevant, providing a measure of serendipitous experiences within the advice process.

Overall serendipity averaged across users and beneficial items could be computed as:

Because the industry evolves, there may be a growing emphasis on refining advice algorithms to deliver recommendations that encompass everything of user preferences, including richer personalization, serendipity, and novelty. Furthermore, advice systems that optimize the balance between these dimensions have also been related to improved user retention metrics and user experience. Ultimately, the goal is to create advice systems that not only cater to users’ known preferences but additionally surprise and delight them with fresh, diverse, and personally relevant recommendations, fostering long-term engagement and satisfaction.

  1. Diversity, Serendipity, Novelty, and Coverage: A Survey and Empirical Evaluation of Beyond-Accuracy Objectives in Recommender Systems
  2. Post Processing Recommender Systems for Diversity
  3. Diversity in recommender systems — A survey
  4. Avoiding congestion in recommender systems
  5. The Definition of Novelty in Advice System
  6. Novelty and Diversity in Recommender Systems: an Information Retrieval approach for evaluation and improvement
  7. Quantifying Availability and Discovery in Recommender Systems via Stochastic Reachability
  8. A recent system-wide diversity measure for recommendations with efficient algorithms
  9. Automatic Evaluation of Advice Systems: Coverage, Novelty and Diversity
  10. Serendipity: Accuracy’s Unpopular Best Friend in Recommenders

1 COMMENT

LEAVE A REPLY

Please enter your comment!
Please enter your name here