Addressing the Accuracy-Bias-Fairness Trade-Off in Recommender Systems via Effective Hyper-parameter tuning: A Pareto Optimality-Based Approach
This is the official implementation of the paper Addressing the Accuracy-Bias-Fairness Trade-Off in Recommender Systems via Effective Hyper-parameter tuning: A Pareto Optimality-Based Approach, under review as full paper at The Web Conf 2023.
This repository is heavily dependent on the framework Elliot, so we suggest you refer to the official GitHub page and documentation.
All graph models are implemented in PyTorch Geometric
using the version 1.10.2
, with CUDA 10.2
and cuDNN 8.0
.
If you have the possibility to install CUDA on your workstation (i.e., 10.2
), you may create the virtual environment with the requirements files we included in the repository, as follows:
# PYTORCH ENVIRONMENT (CUDA 10.2, cuDNN 8.0)
$ python3 -m venv venv_pt
$ source venv_pt/bin/activate
$ pip install --upgrade pip
$ pip install -r requirements_pt.txt
$ pip install torch-scatter torch-sparse torch-cluster torch-spline-conv torch-geometric -f https://data.pyg.org/whl/torch-1.10.0+cu102.html
A more convenient way of running experiments is to instantiate a docker container having CUDA 10.2
already installed.
Make sure you have Docker and NVIDIA Container Toolkit installed on your machine (you may refer to this guide).
Then, you may use the following Docker image to instantiate the container equipped with CUDA 10.2
:
Container Docker with CUDA 10.2
and cuDNN 8.0
(the environment for PyTorch
): link
After the setup of your Docker containers, you may follow the exact same guidelines as scenario #1.
At ./data/
you may find all tsv files for the datasets, i.e., training, validation, and test sets.
To train and evaluate models an all considered metrics, you may run the following command:
$ python -u start_experiments.py --config <dataset_model>
where <dataset_model>
refers to the name of the dataset and model to consider in the current experiment.
You may find all configutation files at ./config_files/<dataset_model>.yml
, where all hyperparameter spaces and the exploration strategies are reported.
Results about calculated metrics are available in the folder ./results/<dataset_name>/performance/
. Specifically, you need to access the tsv file having the following name pattern: rec_cutoff_<cutoff>_relthreshold_0_<datetime-experiment-end>.tsv
.
If you want to calculate, for each metric pair (e.g., Recall vs. APLT), the configuration points which belong (or not) to the Pareto frontier, and reproduce the results illustrated in the paper, you need to use the script pareto.py
.
Open the file, and modify the following lines for your convenience:
- line 188: modify the path to the tsv file where all configurations for a specific model are reported, along with their own metric results (Elliot generates this file when the whole experimental flow is over, you may find it at
./results/performance/
- lines 202-203: decide what to comment/uncomment based on the multi-objective trade-off you are considering
Once the script has been run and it is over, you will end up with a csv file indicating, for each point in the objective space, its coordinates and whether it belongs to the Pareto frontier or not.