On November 16th, FogProtect is represented at the 18th International Conference on Service-Oriented Computing (ICSOC 2020).
The ICSOC is taking place from 14th to 17th of November to create an international virtual environment for researchers, professionals, and industrial practitioners to share groundbreaking research work in the exciting topic of Service-Oriented Computing.
Experts from diverse related fields – business-process management, distributed systems, computer networks, wireless and mobile computing, cloud/edge/fog computing, cyber-physical systems, networking, scientific workflows, services science, data science, management science, software engineering – will carry out presentations, workshops, demonstrations and special sessions with a particular focus on the theme “The Era of Data Science and Artificial Intelligence”.
On the 17th of November, Andreas Metzer, our partner from Paluno – The Ruhr Institute for Software Technology, at the University of Duisburg-Essen will be presenting an accepted paper that acknowledges the research work carried out in FogProtect regarding a cloud use case.
The paper entitled “Feature-Model-Guided Online Reinforcement Learning for Self-Adaptive Services” is written by Andreas Metzger, Clément Quinton, Zoltan Adam Mann; Luciano Baresi, and Klaus Pohl.
Soon you will be able to read the full document at our Publications section. Meet here the abstract:
“A self-adaptive service can maintain its QoS requirements in the presence of dynamic environment changes. To develop a self-adaptive service, service engineers have to create self-adaptation logic encoding when the service should execute which adaptation actions. However, developing self-adaptation logic may be difficult due to design time uncertainty; e.g., anticipating all potential environment changes at design time is in most cases infeasible. Online reinforcement learning addresses design time uncertainty by learning suitable adaptation actions through interactions with the environment at runtime. To learn more about its environment, reinforcement learning has to select actions that were not selected before, which is known as exploration. How exploration happens has an impact on the performance of the learning process. We focus on two problems related to how a service’s adaptation actions are explored:
1. Existing solutions randomly explore adaptation actions and thus may exhibit slow learning if there are many possible adaptation actions to choose from.
2. Existing solutions are unaware of service evolution, and thus may explore new adaptation actions introduced during such evolution rather late. We propose novel exploration strategies that use feature models (from software product line engineering) to guide exploration in the presence of many adaptation actions and in the presence of service evolution. Experimental results for a self-adaptive cloud management service indicate an average speed-up of the learning process of 58.8% in the presence of many adaptation actions, and of 61.3% in the presence of service evolution. The improved learning performance in turn led to an average QoS improvement of 7.8% and 23.7% respectively.”