Events in October–November 2023
MonMonday | TueTuesday | WedWednesday | ThuThursday | FriFriday | SatSaturday | SunSunday |
---|---|---|---|---|---|---|
September 25, 2023
|
September 26, 2023
|
September 27, 2023
|
September 28, 2023
|
September 29, 2023
|
September 30, 2023
|
OctoberOctober 1, 2023 |
October 2, 2023
|
October 3, 2023
|
October 4, 2023
|
October 5, 2023
|
October 6, 2023
|
October 7, 2023
|
October 8, 2023
|
October 9, 2023
|
October 10, 2023
|
October 11, 2023
|
October 12, 2023
|
October 13, 2023
|
October 14, 2023
|
October 15, 2023
|
October 16, 2023
|
October 17, 2023
|
October 18, 2023
|
October 19, 2023
|
October 20, 2023
|
October 21, 2023
|
October 22, 2023
|
October 23, 2023
|
October 24, 2023
|
October 25, 2023
|
October 26, 2023
|
October 27, 2023
|
October 28, 2023
|
October 29, 2023
|
October 30, 2023
|
October 31, 2023
|
NovemberNovember 1, 2023 |
November 2, 2023
|
November 3, 2023
|
November 4, 2023
|
November 5, 2023
|
November 6, 2023
|
November 7, 2023
|
November 8, 2023
|
November 9, 2023
|
November 10, 2023
|
November 11, 2023
|
November 12, 2023
|
November 13, 2023
|
November 14, 2023
|
November 15, 2023
|
November 16, 2023
|
November 17, 2023
|
November 18, 2023
|
November 19, 2023
|
November 20, 2023
|
November 21, 2023
|
November 22, 2023
|
November 23, 2023
|
November 24, 2023
|
November 25, 2023
|
November 26, 2023
|
November 27, 2023
|
November 28, 2023
|
November 29, 2023
|
November 30, 2023
|
DecemberDecember 1, 2023 |
December 2, 2023
|
December 3, 2023
|
- February 7, 2023 @ Bâtiment IMAG (406) -- Seminar Polaris-tt Learning in finite-horizon MDP with UCB (Romain Cravic)
Most of you probably know Markov Decisions Processes (MDP). They are very useful to handle situations where an agent interacts with an environnement that may involve randomness. Concretely, at each time the MDP has a current state and the agent chooses an action : This couple state-action induces a (random) reward and a (random) state transition. If the probability distributions for rewards and transitions are known, at least theoretically, designing optimal behaviors for the agent is easy. What about the case where these distributions are unknown at the early stage of the process ? How to LEARN optimal behaviors efficiently ? A popular way to handle this issue is to use the optimism paradigm, inspired from UCB algorithms designed for stochastic bandits problems. In this talk, I will expose the main ideas of two possible approaches, UCRL algorithm and optimistic Q-learning algorithm, that use optimism to well perform in finite-horizon
- February 21, 2023 @ Bâtiment IMAG (406) -- Seminar Polaris-tt: Decomposition of Normal Form Games - Harmonic, Potential, and Non-Strategic Games (Davide Legacci)
In this talk, we will explore the concept of normal form games and their decomposition into non-strategic, harmonic, and potential games. We will begin by introducing the response graph of a game, which is a visual representation of the strategies available to each player and their corresponding utilities. What dictates the strategic interaction among players is the difference between utilities, rather than the utilities themselves. We will introduce an object that captures this behavior, called deviation flow of the game, and use it to define non-strategic, harmonic, and potential games. Finally, we will discuss the properties of these components.
- March 30, 2023 @ Bâtiment IMAG (amphitheater) -- PhD defense Kimang Khun: Apprentissage par renforcement dans les systèmes dynamiques structurésThèse supervisée par Nicolas GAST et Bruno GAUJAL.