Elsevier

Computer Networks

Volume 53, Issue 10, 14 July 2009, Pages 1712-1721
Computer Networks

Self-adaptive utility-based web session management

https://doi.org/10.1016/j.comnet.2008.08.022Get rights and content

Abstract

In the Internet, where millions of users are a click away from your site, being able to dynamically classify the workload in real time, and predict its short term behavior, is crucial for proper self-management and business efficiency. As workloads vary significantly according to current time of day, season, promotions and linking, it becomes impractical for some ecommerce sites to keep over-dimensioned infrastructures to accommodate the whole load. When server resources are exceeded, session-based admission control systems allow maintaining a high throughput in terms of properly finished sessions and QoS for a limited number of sessions; however, by denying access to excess users, the website looses potential customers.

In the present study we describe the architecture of AUGURES, a system that learns to predict Web user’s intentions for visiting the site as well its resource usage. Predictions are made from information known at the time of their first request and later from navigational clicks. For this purpose we use machine learning techniques and Markov-chain models. The system uses these predictions to automatically shape QoS for the most profitable sessions, predict short-term resource needs, and dynamically provision servers according to the expected revenue and the cost to serve it. We test the AUGURES prototype on access logs from a high-traffic, online travel agency, obtaining promising results.

Introduction

According to the Internet World Stats [19], there are over 1 billion Internet users, and when online, they are a click away from any site. Certain events such as news, world events, or promotions announced in mass media or in the Web can cause a flock of users to visit related websites, creating peak loads in minutes. Furthermore, current Web applications rely on technologies such as XML-based Web services for B2B communication, SSL for security, media streaming, and AJAX for interactivity. While these technologies enhance users’ experience and privacy, they also increase the demand for CPU and resources on the server side. As Web applications become more resource-intensive and the number of potential visitors increases, system overload incidence is also growing along.

Scaling the infrastructure of a website might not be simple; for cost reasons, scalability problems, or because some peaks are infrequent, websites may not be able to adapt rapidly in hardware to user fluctuations. When a server is overloaded, it will typically serve no connection, as connections compete for resources. System administrators might get warnings from resource-monitoring services, but in general they get to the problem when the situation has already occurred, and controlling the arrival of users is out of their reach. To address this issue, session-based admission control systems [3], [6] are used to keep a high throughput in terms of properly finished sessions and QoS for a limited number of sessions. However, by denying access to excess users, the website loses potential revenue from customers.

In this study we propose a new approach to this problem which uses learning techniques applied to past data, to create a model for anonymous Web user behavior in a real, complex website that does experience the overload situations mentioned above. The model is then used to support decisions regarding the allocation of the available resources, based on a utility-related metrics. The learning phase captures the selected features in a model according to a utility goal. As a proof of concept, we have selected the features that make a customer more likely to make a purchase, and therefore more attractive – from the point of view of maximizing revenues – to keep in the system in the case of a severe overload. In most of the article, we take as a metric the number of completed sessions that end in purchase; Section 6 describes other possible metrics and applications. In this sense, we are proposing a per-user adaptive utility-based session management.

We describe the architecture of the prototype we have developed, named AUGURES, and the simulation experiments we have performed on weblogs from the servers of a high-traffic online travel agency, Atrapalo.com. The experiments indicate that using AUGURES to prioritize customer sessions can lead to increased revenue in at least two situations: one, when overload situations occur; that is, the incoming transaction load exceeds the site’s capacity and some sessions will have to be queued, redirected to a static site, or dropped; for this study, these should be mostly non-buying sessions, while we try to admit most buying ones. The second scenario is that in which keeping a server running has a quantifiable cost; in this case, one could try to group buying sessions on a small number of servers, possibly shutting down those other servers that would produce little or no revenue.

Our preliminary experiments [13] showed that the users’ intentions for visiting a site can be predicted to some extent from their navigation clicks, results of previous visits, and other session information. In this context, a revenue-based admission control policy can help to avoid revenue loss due to randomly delaying or dropping excess connections. Defining admission policies based on information generated from user behavior models can contribute to devising cost-effective infrastructures, and seems to have promising applications for resource management for medium to large Web infrastructures.

In this paper we present a method for learning, from the analysis of session logs, how to assign priorities to customers according to some metric – in this study, to their purchasing probability in the current session. Our approach consists in using the Web server log files to develop learning models that make predictions about each class of user future behavior, with the objective of assigning a priority value to every customer based on the expected revenue that s/he will generate, which in our case essentially depends on whether s/he will make a purchase. Our learning methods combines static information (time of access, URL, session ID, referer, among others) and dynamic information (the Web graph of the path followed by the user), in order to make predictions for each incoming Web request.

Section snippets

Related work

In the context of Web workload analysis, there are few published studies based on real e-commerce data, mainly because companies consider HTTP logs as sensitive data. Moreover, most works are based on static content sites, where the studied factors were mainly: file size distributions, which tend to follow a Pareto distribution [1]; and file popularity following Zipf’s Law [1], [8]. Also, works such as [11] have studied logs from real and simulated auction sites and bookstores; there are no

The AUGURES prototype

In this section we describe the architecture of AUGURES, the prototype we have implemented to perform the experiments. AUGURES currently has two subsystems: an offline component (the learner) that takes the historical logfile and produces a predictive model or predictor, and a real-time component, the selector, implemented as a service that runs along with the session manager of the firewall. The selector analyses the incoming requests, runs them through the predictor, and outputs the priority

Results and evaluation

In this section we describe the data, our experiments with the prototype, and discuss the results obtained. We want to remark that our goal was to test a generic approach without fine tuning the experiments with domain knowledge, rather than obtaining the best possible figures for this particular dataset.

The effect of dynamic server activation

In the next experiment we wanted to simulate the benefit of our technique in an environment where not all servers have to be active at all times, but where they can be turned on and shut down dynamically. This possibility is today routinely used in all centers having a reasonable number of in-site servers: in this case, shutting down unnecessary servers results in immediate savings in power (both for the server itself and for cooling), hence lower costs. Also, dynamic provisioning of external

Other potential applications

This section describes other potential applications for the techniques described in the article. We summarize as using machine learning techniques to characterize individual anonymous Web sessions in real time from available session information and past data.

A first approach to extended the experiments presented in this article, is to use other metrics than if the session will end up in a purchase, to prioritize it. Two that we plan to investigate in the immediate future are:

  • Expected purchase

Conclusions and future work

Websites might become overloaded by certain events such as news events or promotions, as they can potentially reach millions of users. When a peak situation occurs most infrastructures become stalled and throughput is reduced. To prevent this, load admission control mechanisms are used to allow only a certain number of sessions; but as they do not differentiate between users, users with intentions to purchase might be denied access. As a proof of concept, we have taken data from a high-traffic

Acknowledgements

Thanks to the reviewers for interesting comments and contributions, to Atrapalo.com, who provided the experiment datasets and domain knowledge.

This research work is supported by the Ministry of Science and Technology of Spain under Contract TIN2007-60625.

R. Gavaldà is partially supported by the 6th Framework Program of EU through the integrated project DELIS (#001907), by the EU PASCAL and PASCAL2 Networks of Excellence, and by the DGICYT MOISES-TA project, TIN2005-08832-C03.

For additional

Nicolas Poggi received his IT Engineering degree with a minor in Business Administration, at the American University (UA), 2005. He is now a Ph.D. candidate at the Technical University of Catalonia in the Computer Architecture Department, DAC, where he has obtained the Diploma of Advanced Studies (DEA). He is part of the Barcelona eDragon Research Group, integrated with the High Performance Computing Group at DAC. Since 1999, he has been working as a consultant for ISPs and eCommerce projects,

References (20)

  • J. Guitart et al.

    Designing an overload control strategy for secure e-commerce applications

    Comput. Networks

    (2007)
  • V. Almeida, M. Crovella, A. Bestavros, A. Oliveira, Characterizing reference locality in the WWW, in: Proceedings of...
  • D. Bonino, F. Corno, G. Squillero, A real-time evolutionary algorithm for Web prediction, in: Proceedings of the...
  • L. Cherkasova et al.

    Session-based admission control: a mechanism for peak load management of commercial Web sites

    IEEE Trans. Comput.

    (2002)
  • G. Cheliotis, C. Kenyon, Autonomic economics: a blueprint for selfmanaged systems, IJCAI Workshop on AI and Autonomic...
  • M. Deshpande et al.

    Selective Markov models for predicting web page accesses

    ACM Trans. Internet Technol. (TOIT)

    (2004)
  • B. Lan, S. Bressan, B.C. Ooi, K. Tan, Rule-assisted prefetching in Web-server caching, in: Proceedings of theACM...
  • M. Levene et al.

    Zipf’s law for web surfers

    Knowl. Inform. Syst.

    (2001)
  • M.L. Littman, T. Nguyen, H. Hirsh, A model of cost-sensitive fault mediation, IJCAI Workshop on AI and Autonomic...
  • D.A. Menascé, V.A.F. Almeida, R. Fonseca, M.A. Mendes, A methodology for workload characterization of e-commerce sites,...
There are more references available in the full text version of this article.

Cited by (26)

  • SLA negotiation and enforcement policies for revenue maximization and client classification in cloud providers

    2014, Future Generation Computer Systems
    Citation Excerpt :

    They maximize the revenue of e-commerce servers by dynamically calculating Customer Behavior Model Graphs and prioritizing the workloads of the users that are willing to spend more money. Poggi et al. [24] introduce a similar approach, in which QoS for user sessions are shaped based on the commercial intention of the users.

  • Adaptive admission control algorithm in a QoS-aware Web system

    2012, Information Sciences
    Citation Excerpt :

    If the arrival rate exceeds a limit, then the admission control decisions are taken each time a new session arrives to the system. A self-adaptive architecture that learns user behaviour and predictes the purchase probability of Web sessions is described by Poggi et al. [39]. A method that does not require any modification in either in the Operative System (OS) or the Web software (Web server, application server or database) is presented in [18] by Elnikety et al.

  • Process mining approach to analyze user navigation behavior of a news website

    2021, ACM International Conference Proceeding Series
View all citing articles on Scopus

Nicolas Poggi received his IT Engineering degree with a minor in Business Administration, at the American University (UA), 2005. He is now a Ph.D. candidate at the Technical University of Catalonia in the Computer Architecture Department, DAC, where he has obtained the Diploma of Advanced Studies (DEA). He is part of the Barcelona eDragon Research Group, integrated with the High Performance Computing Group at DAC. Since 1999, he has been working as a consultant for ISPs and eCommerce projects, also, holds a technology transfer research scholarship.

Toni Moreno obtained his B.Sc. in Electrical Engineering (1999) and M.Sc. degrees in Electrical Engineering (2003) and Industrial Engineering (2003) from the Universitat Politècnica de Catalunya, UPC. Since 1998 he has held several positions in the IT industry, mainly as a project manager. Between 2004 and 2007, he was also an adjunct lecturer at the Management Department of the Universitat Politècnica de Catalunya, UPC. He is currently a Doctoral Candidate at the Operations and Information Management Department of The Wharton School (University of Pennsylvania).

Josep L. Berral obtained his M.Sc. degree (2007) in Information Technologies from Universitat Politècnica de Catalunya, UPC. He is now attending at the Ph.D. courses on Computer Architecture and Technology, at the Computer Architecture Department. Also he is working on applications of Machine Learning to Autonomic Computing with the eDragon Research Group, in the UPC High Performance Computing group.

Ricard Gavaldà received his Degree (1987) and Ph.D. (1992) in Computer Science from Universitat Politècnica de Catalunya, UPC. He has had a permanent position at the Departament de Llenguatges i Sistemes Informàtics of UPC since 1993, where he became full professor in 2008. His main research field was initially Computational Complexity Theory, and has gradually evolved towards Computational Learning Theory and algorithmic aspects of Machine Learning and Data Mining. He has recently started working on applications of Machine Learning to Autonomic Computing.

Jordi Torres has a Masters degree in Computer Science from the Technical University of Catalonia (UPC, 1988) and also holds a Ph.D. in Computer Science from the same institution (UPC, Best UPC Computer Science Thesis Award, 1993). Currently he is a full professor in the Computer Architecture Department (DAC) at UPC. He leads the Barcelona eDragon Research Group, which is integrated with the High Performance Computing Group at DAC, and is a Manager for the Autonomic Systems and eBusiness Platforms research line in Barcelona Supercomputing Center (BSC). At this moment in time his research is centered on maximizing the return on investment of IT resources, thereby making them more sustainable, and focuses on areas such as Autonomic Computing, Parallel and Distributed systems, Performance Modeling, Machine Learning and Cloud Computing.

1

Present address: Department of Operations and Information Management of The Wharton School, University of Pennsylvania.

View full text