SEMINARS

Events in October–November 2023

Monday Tuesday Wednesday Thursday Friday Saturday Sunday
September 25, 2023
September 26, 2023
September 27, 2023
September 28, 2023
September 29, 2023
September 30, 2023

October

October 1, 2023
October 2, 2023
October 3, 2023
October 4, 2023
October 5, 2023
October 6, 2023
October 7, 2023
October 8, 2023
October 9, 2023
October 10, 2023
October 11, 2023
October 12, 2023
October 13, 2023
October 14, 2023
October 15, 2023
October 16, 2023
October 17, 2023
October 18, 2023
October 19, 2023
October 20, 2023
October 21, 2023
October 22, 2023
October 23, 2023
October 24, 2023
October 25, 2023
October 26, 2023
October 27, 2023
October 28, 2023
October 29, 2023
October 30, 2023
October 31, 2023

November

November 1, 2023
November 2, 2023
November 3, 2023
November 4, 2023
November 5, 2023
November 6, 2023
November 7, 2023
November 8, 2023
November 9, 2023
November 10, 2023
November 11, 2023
November 12, 2023
November 13, 2023
November 14, 2023
November 15, 2023
November 16, 2023
November 17, 2023
November 18, 2023
November 19, 2023
November 20, 2023
November 21, 2023
November 22, 2023
November 23, 2023
November 24, 2023
November 25, 2023
November 26, 2023
November 27, 2023
November 28, 2023
November 29, 2023
November 30, 2023

December

December 1, 2023
December 2, 2023
December 3, 2023
  • February 7, 2023 @ Bâtiment IMAG (406) -- Seminar Polaris-tt Learning in finite-horizon MDP with UCB (Romain Cravic)

    Most of you probably know Markov Decisions Processes (MDP). They are very useful to handle situations where an agent interacts with an environnement that may involve randomness. Concretely, at each time the MDP has a current state and the agent chooses an action : This couple state-action induces a (random) reward and a (random) state transition.  If the probability distributions for rewards and transitions are known, at least theoretically, designing optimal behaviors for the agent is easy. What about the case where these distributions are unknown at the early stage of the process ? How to LEARN optimal behaviors efficiently ? A popular way to handle this issue is to use the optimism paradigm, inspired from UCB algorithms designed for stochastic bandits problems. In this talk, I will expose the main ideas of two possible approaches, UCRL algorithm and optimistic Q-learning algorithm,  that use optimism to well perform in finite-horizon

  • February 21, 2023 @ Bâtiment IMAG (406) -- Seminar Polaris-tt: Decomposition of Normal Form Games - Harmonic, Potential, and Non-Strategic Games (Davide Legacci)

    In this talk, we will explore the concept of normal form games and their decomposition into non-strategic, harmonic, and potential games. We will begin by introducing the response graph of a game, which is a visual representation of the strategies available to each player and their corresponding utilities. What dictates the strategic interaction among players is the difference between utilities, rather than the utilities themselves. We will introduce an object that captures this behavior, called deviation flow of the game, and use it to define non-strategic, harmonic, and potential games. Finally, we will discuss the properties of these components.

  • March 30, 2023 @ Bâtiment IMAG (amphitheater) -- PhD defense Kimang Khun: Apprentissage par renforcement dans les systèmes dynamiques structurés
    Thèse supervisée par Nicolas GAST et Bruno GAUJAL.

Comments are closed.