In this article we demonstrate how to use a Field-Aware Factorization Machine to recommend hotels on the Las Vegas Strip, based on the traveler type (solo, family, business, …) and the season. We’ll use ML.NET for the Machine Learning stuff, OxyPlot for visualization, and a UWP app to host it all. This is how the sample page looks like:
This page looks pretty much like the one in our previous article on recommendation in Machine Learning. That article also recommended hotels – but since we were using Matrix Factorization we could only use one feature to base the recommendation on (i.c. traveler type). In the previous article, the prediction model was defined by preparing the two feature columns (TravelerType and Hotel) and passing these to a recommendation algorithm, like this:
var pipeline = _mlContext.Transforms.Conversion.MapValueToKey("Hotel") .Append(_mlContext.Transforms.Conversion.MapValueToKey("TravelerType")) .Append(_mlContext.Recommendation().Trainers.MatrixFactorization( labelColumnName: "Label", matrixColumnIndexColumnName: "Hotel", matrixRowIndexColumnName: "TravelerType")) .Append(_mlContext.Transforms.Conversion.MapKeyToValue("Hotel")) .Append(_mlContext.Transforms.Conversion.MapKeyToValue("TravelerType"));
The corresponding pipeline in this article would look very similar:
var pipeline = _mlContext.Transforms.Categorical.OneHotEncoding("TravelerTypeOneHot", "TravelerType") .Append(_mlContext.Transforms.Categorical.OneHotEncoding("HotelOneHot", "Hotel")) .Append(_mlContext.Transforms.Concatenate("Features", "TravelerTypeOneHot", "HotelOneHot")) .Append(_mlContext.BinaryClassification.Trainers.FieldAwareFactorizationMachine(new string[] { "Features" }));
The advantage of a Field-Aware Factorization Machine over Matrix Factorization is that you’re not limited to two features. This allows you to provide much better recommendations. Hotels could be recommended on a combination of traveler type, season, country of origin, etc.
Field-Aware Factorization in Machine Learning
A Field-Aware Factorization Machine (FFM) is a recommendation algorithm that is specialized in deriving knowledge from large and sparse datasets. It recognizes feature conjunctions in feature vectors. This is particularly useful in Click-Through Rate prediction (CTR). Check this article for an introduction and a comparison to other algorithms.
The FFM algorithm takes a vector of numerical features as input. When we’re dealing with categorical data (countries, months, traveler types, …) we need to transform these into numbers. In the context of FFM the right approach is to use One-Hot Encoding, which boils down to pivoting feature values into separate columns. There’s an excellent beginners guide to one-hot encoding right here, but this illustration from Chris Albon’s Machine Learning Flashcards says it all:
The one-hot encoding transformation extends the schema with a lot of new features, sparsely filled. That exactly how FFM likes it…
Field-Aware Factorization in ML.NET
With the FieldAwareFactorizationMachineTrainer and the OneHotEncodingTransformer, ML.NET has all the ingredients to implement a Field-Aware Factorization Machine scenario.
Let’s write some code
Here’s how a typical Machine Learning use case looks like:
Raw data is collected, then cleaned up, completed, and transformed to fit the algorithm. Part of the data is used to train the model, another part is used to evaluate it. When the model passed all tests, it is persisted for consumption.
The sample app uses a light-weight MVVM architecture. The beef of the ML.NET manipulation is in this Model class.
Getting the Raw Data
The raw data in the sample app comes from a comma-separated flat file with Trip Advisor reviews from 2015 for hotels on the Las Vegas Strip:
Preparing the Data
While we read the data, we already start the preparation. FFM is a binary classification algorithm, so we need to transform the rating (0-5) to a Boolean value (“recommended or not”). Here’s the input structure:
public class FfmRecommendationData { public bool Label; public string TravelerType; public string Season; public string Hotel; }
The predicted label will also be a Boolean, and the algorithm also provides the corresponding probability. Here’s the model’s output structure:
public class FfmRecommendationPrediction { public bool PredictedLabel; public float Probability; public string TravelerType; public string Season; public string Hotel; }
Before we read all the data in an IDataView, we need to create a MLContext instance:
private MLContext _mlContext = new MLContext(seed: null);
While we read the data, we transform the score (0 to 5) into a Boolean by comparing it to a threshold value. We used 3 as the threshold, since the general ratings in the dataset are pretty high – which is probably why these were allowed to make public.
With the LoadFromEnumerable() method we transform it into an IDataView:
private IDataView _allData; private ITransformer _model; public IEnumerable<FfmRecommendationData> Load(string trainingDataPath) { // Populating an IDataView from an IEnumerable. var data = File.ReadAllLines(trainingDataPath) .Skip(1) .Select(x => x.Split(';')) .Select(x => new FfmRecommendationData { Label = double.Parse(x[4]) > _ratingThreshold, Season = x[5], TravelerType = x[6], Hotel = x[13] }); _allData = _mlContext.Data.LoadFromEnumerable(data); // Just 'return data;' would also do the trick... return _mlContext.Data.CreateEnumerable<FfmRecommendationData>(_allData, reuseRowObject: false); }
For the next step in preparing the data we will rely on some ML.NET transformations. First we apply a OneHotEncoding transformation to all the features, to transform the categorical data into numbers. Then all features are combined into a vector with a call to Concatenate(). The model building pipeline is then completed with a FieldAwareFactorizationMachine:
var pipeline = _mlContext.Transforms.Categorical.OneHotEncoding("TravelerTypeOneHot", "TravelerType") .Append(_mlContext.Transforms.Categorical.OneHotEncoding("SeasonOneHot", "Season")) .Append(_mlContext.Transforms.Categorical.OneHotEncoding("HotelOneHot", "Hotel")) .Append(_mlContext.Transforms.Concatenate("Features", "TravelerTypeOneHot", "SeasonOneHot", "HotelOneHot")) .Append(_mlContext.BinaryClassification.Trainers.FieldAwareFactorizationMachine(new string[] { "Features" }));
Training the Model
To train the model, we send 450 randomly chosen rows from the dataset to it. The rows are selected with some methods from the DataOperationsCatalog: first we rearrange the dataset with ShuffleRows(), then we pick some rows with TakeRows():
var trainingData = _mlContext.Data.ShuffleRows(_allData); trainingData = _mlContext.Data.TakeRows(trainingData, 450);
After training the model, we’ll create a strongly typed PredictionEngine from it, for individual recommendations. So we declared a field to host this:
private PredictionEngine<FfmRecommendationData, FfmRecommendationPrediction> _predictionEngine;
The model is created and trained with a call to Fit(), and then we create the prediction engine from it with CreatePredictionEngine():
_model = pipeline.Fit(trainingData); _predictionEngine = _mlContext.Model.CreatePredictionEngine<FfmRecommendationData, FfmRecommendationPrediction>(_model);
Testing the Model
To test the model we send another 100 random rows from the dataset to it and call the Transform() to generate the predictions. A call to the Evaluate() method in the BinaryClassificationCatalog will compare the predicted labels to the original ones:
public CalibratedBinaryClassificationMetrics Evaluate(string testDataPath) { var testData = _mlContext.Data.ShuffleRows(_allData); testData = _mlContext.Data.TakeRows(testData, 100); var scoredData = _model.Transform(testData); var metrics = _mlContext.BinaryClassification.Evaluate( data: scoredData, labelColumnName: "Label", scoreColumnName: "Probability", predictedLabelColumnName: "PredictedLabel"); // Place a breakpoint here to inspect the quality metrics. return metrics; }
The result of the evaluation is an instance of CalibratedBinaryClassificationMetrics with useful statistics such as accuracy, entropy, recall, and F1-score:
Persisting the Model
When you’re happy with the model’s quality, then you can serialize and persist it for later use with a call to the Save() method from the ModelOperationsCatalog:
public void Save(string modelName) { var storageFolder = ApplicationData.Current.LocalFolder; string modelPath = Path.Combine(storageFolder.Path, modelName); _mlContext.Model.Save(_model, inputSchema: null, filePath: modelPath); }
Consuming the Model
There are two ways to consume the model. With a call to Transform() you can generate predictions or recommendations for a group of input records. The resulting IDataView can be transformed to a list of prediction records with CreateEnumerable():
public IEnumerable<FfmRecommendationPrediction> Predict(IEnumerable<FfmRecommendationData> recommendationData) { // Group prediction var data = _mlContext.Data.LoadFromEnumerable(recommendationData); var predictions = _model.Transform(data); return _mlContext.Data.CreateEnumerable<FfmRecommendationPrediction>(predictions, reuseRowObject: false); }
The strongly typed PredictionEngine that we created after training the model, can be used for single recommendation. Its Predict() method runs the prediction pipeline for a single input record:
public FfmRecommendationPrediction Predict(FfmRecommendationData recommendationData) { // Single prediction return _predictionEngine.Predict(recommendationData); }
When the predicted label is false, the model does not recommend the hotel/travelertype/season combination. In that case we reverse the displayed probability (the score) so that its values become a range from –1 (strongly discouraged) to +1 (strongly recommended). This is done in the MVVM View (the code-behind in the page):
var result = await ViewModel.Predict(recommendationData); if (!result.PredictedLabel) { // Bring to a range from -1 (highly discouraged) to +1 (highly recommended). result.Probability = -result.Probability; }
The Model in Action
Here’s the FFM Recommendation page from the sample app again:
When the model is ready for operation, the combo boxes are populated and unlocked. When you change the traveler type or the season, a group prediction is done for all the hotels in the data set. Its result is displayed on the left in a horizontal bar chart. We’ll not diving into its details, since it’s basically the same diagram as the one in the previous article (we’re just displaying the probability (-1 to +1) instead of the predicted rating (0 to 5).
When you select a hotel in the combo box in the bottom left corner, a single prediction is made, and the result is displayed next to it. The diagram for the group predictions only displays recommended hotels, but in the single prediction you can pick your own hotel. That’s why we decided to display a negative probability for negative advice.
If you want to run this scenario yourself, feel free to download the sample app. Its source lives here on GitHub.
Enjoy!