Recommendation Systems

Rec The Trail

Posted by Jade on October 29, 2016

This is part two of three part series about my capstone project, Rec The Trail, that I completed for the Galvanize Data Science Immersive program in Seattle, WA in 2016.

Recommender Systems

If you’ve spent any amount of time on the internet, you’ve most likely come across recommendation systems. Netflix telling you which movies to watch, Amazon telling you what other things you might want to buy, CNN telling you which articles are similar to the one that you just read. So they are all over the place, but how do we build them?

The short answer, it depends on what data you have available.

The Turi (recently bought by Apple) blog offers a great breakdown of the different recommender systems and how to decide which ones to use, but I’ll give a brief breakdown here as well. The type of data you have- explicit, implicit or item content- will dictate what kind of recommender you use. Explicit data means that you have a matrix of actual rating data and a recommender system using this data is trying to predict which other items a user might rate highly, given the ratings of similar users. The “Users who liked this product, also liked these products” recommenders. With implicit data, there is no rating data, just user and item data. An example of this might be if you have purchasing data, but no ratings. A recommender system would then make recommendations off of which items have seen similar interactions between users. The “Users who bought this item also bought these items” recommenders. Lastly, in the case where you have no user data, you can use an item content similarity recommender. These recommenders base their recommendations off items that most similar to the one that the user liked. These are often the easiest to make and can most effectively deal with the infamous “cold start” problem, but are not always applicable. For example, if a user has bought a tent at REI, you would not want the REI recommendation systems to send you more tent recommendations, but instead items that other users think went with their tents (stoves, sleeping bags, disco balls…).

Recommend a Trail for Me

So what did I use for Rec the Trail? If you read my post on sentiment analysis, you’ll know that I created my own user rating matrix from trip report text. Therefore, I had explicit data. Because of that, I was able to build a ranking factorization recommender in Turi GraphLab.

Great, easy enough… Almost too easy.

Yes. Yes it was. Turns out, if you ask the recommender for a recommendation for someone with no existing ratings, it will pick the same, most popular hikes- with no degree of personalization. This is caused by what is known as the cold start problem.

The Cold Start Problem

A common problem in recommender systems, a cold start describes an instances of a new user in a system on which little personal preference or rating data is available. An example of this would be a new user to Netflix. Since the recommender system knows very little about what kind of movies the new user likes, there will be little to no personalization until ratings are provided. As a result, you will most likely get recommendations to watch only the most popular movies on Netflix. Even once you provide one or two ratings, there is such a large corpus of movies and genres that it will still be difficult for the system to pick out the nuances about what makes a movie appealing to a certain person.

Since this is by no means a novel problem in recommender system performance, there are many available solutions. One of which is item clustering and item similarity. In the Netflix example, movies can be clustered based on genre, year, actors, tone, so on. So when you visit or rate an item/movie/product, it will show you items that are similar to the one that you visited. So in the case of my recommender, if someone inputs a hike that they like but don’t have any previous “ratings“, I just show them hikes that are most similar to the one that they input.

But now the question becomes, most similar based on what criteria? When I scraped the WTA website, I scraped the following metadata:

  • Elevation gain
  • Total length
  • Latitude and longitude coordinates- which I then used to get the drive time from Seattle.
  • Features of the hike such as waterfalls, lakes, child friendly, etc
  • Region
  • Overall star rating
  • Number of trip reports- which I believe is a proxy for hike popularity

Initially, I tested the item similarity recommender with all of the features having equal weights. After eyeballing the results of the recommender, I was finding that the hikes recommended were actually not that similar, or at least weren’t giving people the kind of hikes that they might expect. For example, when I prompted the recommender using Mount Pilchuck, a rather strenuous hike, it was returning flat loop hikes through Discovery Park- a state park in the middle of Seattle. A lovely hike, but doubtfully similar to Mount Pilchuck. Because of this, I wanted to do a little recon on what features of hike are actually most important to people.

In order to determine which features people care about the most, I ran a quick random forest model with average star rating as the target and all of the other metadata as the features in order to determine feature importance. I then used the fraction of samples affected as the weights for the features in the item content similarity recommender.

Feature Importance in Predicting Average Star Rating

After eyeballing the results from these recommenders (the best way I could find to evaluate the models), I determined that this was a good way to deal with the cold start problems and generally gave people recommendations that made sense.

Give it a try and let me know what you think!

Resources

  1. Turi Machine Learning Platform User Guide. (n.d.). Retrieved August 09, 2016, from https://turi.com/learn/userguide/recommender/choosing-a-model.html
  2. Debnath, S., Ganguly, N., & Mitra, P. (2008). Feature weighting in content based recommendation system using social network analysis. Proceeding of the 17th International Conference on World Wide Web – WWW ’08. doi:10.1145/1367497.1367646
  3. Sobhanam, H., & Mariappan, A. K. (2013). Addressing cold start problem in recommender systems using association rules and clustering technique. 2013 International Conference on Computer Communication and Informatics. doi:10.1109/iccci.2013.6466121