Main investigators:
- Isabelle Chrisment, Inria RESIST
- Prof. Hans Dieter Schotten, DFKI (German Research Center for Artificial Intelligence), Germany
- Dr. Daishi Kondo, Osaka Prefecture University, Japan
Objectives:
ML algorithms suffer from their complexity which results into highly-customized techniques (centric to a use case and even a dataset). Most precisely, they suffer from their difficult configurations with many hyper-parameters to tune or also the algorithm to be chosen. Furthermore, over-fitting during the learning phase prevents the model to be robust against noisy data or generalized against new data, i.e. new type of attackers’ action. Therefore, our main objective is to make robust ML techniques when they will face new types of attacks or when deployed within new environment despite the lack of large and comprehensive datasets.
The auto-configuration of a ML algorithm or the selection of the right algorithm to solve a problem can be automated with hyper-parameter optimisation techniques. Generating additional synthetic data is another alternative to help to learn a more general model. Very recently, federated learning empowers collaborative approaches where several distributed agents learn all together without sharing their data.
All theses techniques are theoretically valid in our context but they have been developed and tested in the context of imagery. It is still uncertain what gain can be expected from them for cyber-security. Our objectives are the investigation, the adaptation of theses techniques (parameter optimization, dataset augmentation, federated learning) and their application to predict attacks.
Furthermore, there are inherent problems that come from these techniques adding complexity on top of ML applications: optimisation of parameters also needs to be configured and is highly computational and data augmentation techniques require to define data transformation functions, which are partially context-specific.