Abstract:Network operators are interested in continuously monitoring the satisfaction of their customers to minimise the churn rate: however, collecting user feedbacks through surveys is a cumbersome task. In this work we explore the possibility of predicting the long-term user satisfaction relative to network coverage and video streaming starting from user-side network measurements only. We leverage country-wide datasets to engineer features which are then used to train several machine learning models. The obtained re… Show more
“…Many authors during the last decade investigated and evaluated the feasibility of predicting both short-term [7], [3] and long-term [8], [2], [9] customers QoE relative to different network services. Short-term QoE concerns individual and time-limited sessions in which users are instructed to use a service (e.g., watching a video content on YouTube) under controlled network environments and are later asked about the quality of their experience.…”
Section: Related Workmentioning
confidence: 99%
“…We pre-process the dataset by considering only the users who replied to QoE feedbacks (i.e., whose identifiers are present in the Q c or Q v datasets), resulting in about 2k users for each scenario. According to our previous study [9], we restrict our attention to a subset of features (10 for network coverage and 14 for video streaming), summarized in Table I. Regardless of the scenario, features are computed independently for each user, considering different periods of d days before the date of the survey.…”
Section: B Crowdsourcing Datasetmentioning
confidence: 99%
“…The variable d relates to the so-called user memory, i.e., for how long a disservice is able to impact the response to a survey. Similarly to what done in [9], we do not use a single value of d, but we set its value to 30, 60 and 90 days before the survey response, each time computing the corresponding subset of features. 1 Finally, we concatenate the three subsets, obtaining a total of 30 and 42 features to be used for each user in network coverage and video streaming scenarios, respectively.…”
Section: B Crowdsourcing Datasetmentioning
confidence: 99%
“…At the end of this process, we obtain 47 features to be used for both network coverage and video streaming QoE prediction. Similarly to what done in [9], we apply to the computed features a log-like transformation, to make their statistical distributions more similar to Gaussian. Finally, features are standardized to mean and variance, a preprocess which benefits ML methods working with normally distributed inputs.…”
Section: Network Kpis Datasetmentioning
confidence: 99%
“…The problem we consider has the form of a binary classification problem, such that each user can be classified as either satisfied (class 0, negative case) or unsatisfied (class 1, positive case or alarm). We take therefore a supervised learning approach: among the several available ML classifiers to be trained we select the Random Forest algorithm, which is widely known to perform well in general and has been successfully applied in the past for similar problems [3], [9]. To compare the two approaches to QoE prediction we proceed as it follows: 1) First, we select the subset of users who appear in both datasets, i.e.…”
Section: Experiments a Training The Classifiersmentioning
Monitoring the Quality of Experience (QoE) of the customer base is a key task for Mobile Network Operators (MNOs), and it is generally performed by collecting users feedbacks through directed surveys. When such feedbacks are few in number, a MNO may predict the users QoE starting from objective network measurements, gathered directly from the users equipments through crowdsourcing. In this work, we compare such a traditional approach with a different one, where the data used for predicting the users QoE is gathered directly at the network access, using Key Performance Indicators (KPI) available on each base station. Although such KPIs are aggregated by design (i.e., they refer to the distribution of a population of users rather than to a single individual), we show through experiments with a country-wide dataset that their predictive power is comparable and in some cases superior than the one of crowdsourcing. Such a result is particularly attractive for MNOs, since network KPIs are generally much easily obtainable than crowdsourcing data.
“…Many authors during the last decade investigated and evaluated the feasibility of predicting both short-term [7], [3] and long-term [8], [2], [9] customers QoE relative to different network services. Short-term QoE concerns individual and time-limited sessions in which users are instructed to use a service (e.g., watching a video content on YouTube) under controlled network environments and are later asked about the quality of their experience.…”
Section: Related Workmentioning
confidence: 99%
“…We pre-process the dataset by considering only the users who replied to QoE feedbacks (i.e., whose identifiers are present in the Q c or Q v datasets), resulting in about 2k users for each scenario. According to our previous study [9], we restrict our attention to a subset of features (10 for network coverage and 14 for video streaming), summarized in Table I. Regardless of the scenario, features are computed independently for each user, considering different periods of d days before the date of the survey.…”
Section: B Crowdsourcing Datasetmentioning
confidence: 99%
“…The variable d relates to the so-called user memory, i.e., for how long a disservice is able to impact the response to a survey. Similarly to what done in [9], we do not use a single value of d, but we set its value to 30, 60 and 90 days before the survey response, each time computing the corresponding subset of features. 1 Finally, we concatenate the three subsets, obtaining a total of 30 and 42 features to be used for each user in network coverage and video streaming scenarios, respectively.…”
Section: B Crowdsourcing Datasetmentioning
confidence: 99%
“…At the end of this process, we obtain 47 features to be used for both network coverage and video streaming QoE prediction. Similarly to what done in [9], we apply to the computed features a log-like transformation, to make their statistical distributions more similar to Gaussian. Finally, features are standardized to mean and variance, a preprocess which benefits ML methods working with normally distributed inputs.…”
Section: Network Kpis Datasetmentioning
confidence: 99%
“…The problem we consider has the form of a binary classification problem, such that each user can be classified as either satisfied (class 0, negative case) or unsatisfied (class 1, positive case or alarm). We take therefore a supervised learning approach: among the several available ML classifiers to be trained we select the Random Forest algorithm, which is widely known to perform well in general and has been successfully applied in the past for similar problems [3], [9]. To compare the two approaches to QoE prediction we proceed as it follows: 1) First, we select the subset of users who appear in both datasets, i.e.…”
Section: Experiments a Training The Classifiersmentioning
Monitoring the Quality of Experience (QoE) of the customer base is a key task for Mobile Network Operators (MNOs), and it is generally performed by collecting users feedbacks through directed surveys. When such feedbacks are few in number, a MNO may predict the users QoE starting from objective network measurements, gathered directly from the users equipments through crowdsourcing. In this work, we compare such a traditional approach with a different one, where the data used for predicting the users QoE is gathered directly at the network access, using Key Performance Indicators (KPI) available on each base station. Although such KPIs are aggregated by design (i.e., they refer to the distribution of a population of users rather than to a single individual), we show through experiments with a country-wide dataset that their predictive power is comparable and in some cases superior than the one of crowdsourcing. Such a result is particularly attractive for MNOs, since network KPIs are generally much easily obtainable than crowdsourcing data.
Network operators need to continuosly upgrade their infrastructures in order to keep their customer satisfaction levels high. Crowdsourcing-based approaches are generally adopted, where customers are directly asked to answer surveys about their user experience. Since the number of collaborative users is generally low, network operators rely on Machine Learning models to predict the satisfaction levels/QoE of the users rather than directly measuring it through surveys. Finally, combining the true/predicted user satisfaction levels with information on each user mobility (e.g, which network sites each user has visited and for how long), an operator may reveal critical areas in the networks and drive/prioritize investments properly. In this work, we propose an empirical framework tailored to assess the quality of the detection of under-performing cells starting from subjective user experience grades. The framework allows to simulate diverse networking scenarios, where a network characterized by a small set of under-performing cells is visited by heterogeneous users moving through it according to realistic mobility models. The framework simulates both the processes of satisfaction surveys delivery and users satisfaction prediction, considering different delivery strategies and evaluating prediction algorithms characterized by different prediction performance. We use the simulation framework to test empirically the performance of under-performing sites detection in general scenarios characterized by different users density and mobility models to obtain insights which are generalizable and that provide interesting guidelines for network operators.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.