Events in March–April 2024
MonMonday | TueTuesday | WedWednesday | ThuThursday | FriFriday | SatSaturday | SunSunday |
---|---|---|---|---|---|---|
February 26, 2024
|
February 27, 2024
|
February 28, 2024
|
February 29, 2024
|
MarchMarch 1, 2024 |
March 2, 2024
|
March 3, 2024
|
March 4, 2024
|
March 5, 2024
|
March 6, 2024
|
March 7, 2024
|
March 8, 2024
|
March 9, 2024
|
March 10, 2024
|
March 11, 2024
|
March 12, 2024
|
March 13, 2024
|
March 14, 2024
|
March 15, 2024
|
March 16, 2024
|
March 17, 2024
|
March 18, 2024
|
March 19, 2024
|
March 20, 2024
|
March 21, 2024
|
March 22, 2024
|
March 23, 2024
|
March 24, 2024
|
March 25, 2024
|
March 26, 2024
|
March 27, 2024
|
March 28, 2024
|
March 29, 2024
|
March 30, 2024
|
March 31, 2024
|
AprilApril 1, 2024 |
April 2, 2024
|
April 3, 2024
|
April 4, 2024
|
April 5, 2024
|
April 6, 2024
|
April 7, 2024
|
April 8, 2024
|
April 9, 2024
|
April 10, 2024
|
April 11, 2024
|
April 12, 2024
|
April 13, 2024
|
April 14, 2024
|
April 15, 2024
|
April 16, 2024
|
April 17, 2024
|
April 18, 2024
|
April 19, 2024
|
April 20, 2024
|
April 21, 2024
|
April 22, 2024
|
April 23, 2024
|
April 24, 2024
|
April 25, 2024
|
April 26, 2024
|
April 27, 2024
|
April 28, 2024
|
April 29, 2024
|
April 30, 2024
|
MayMay 1, 2024 |
May 2, 2024
|
May 3, 2024
|
May 4, 2024
|
May 5, 2024
|
- April 3, 2024 @ Bâtiment IMAG (442) -- [Seminar] Victor Boone
Who: Victor Boone
When: Wednesday, April 3, 14:00-15:00
Where: 447
What: Learning MDPs with Extended Bellman Operators
More: Efficiently learning Markov Decision Processes (MDPs) is difficult. When facing an unknown environment, where is the adequate limit between repeating actions that have shown their efficiency in the past (exploitation of your knowledge) and testing alternatives that may actually be better than what you currently believe (exploration of the environment)? To bypass this dilemma, a well-known solution is the "optimism-in-face-of-uncertainty" principle: Think of the score of an action as being the largest that is statistically plausible.
The exploration-exploitation dilemma then becomes the problem of tuning optimism. In this talk, I will explain how optimism in MDPs can be all rephrased using a single operator, embedding all the uncertainty in your environment within a single MDP. This is a story about "extended Bellman operators" and "extended MDPs", and about how one can achieve minimax optimal regret using this machinery.
- April 11, 2024 @ Bâtiment IMAG (442) -- [Seminar] Charles Arnal
Who: Charles Arnal
When: Thursday, April 11, 14:00-15:00
Where: 442
What: Mode Estimation with Partial Feedback
More: The combination of lightly supervised pre-training and online fine-tuning has played a key role in recent AI developments. These new learning pipelines call for new theoretical frameworks. In this paper, we formalize core aspects of weakly supervised and active learning with a simple problem: the estimation of the mode of a distribution using partial feedback. We show how entropy coding allows for optimal information acquisition from partial feedback, develop coarse sufficient statistics for mode identification, and adapt bandit algorithms to our new setting. Finally, we combine those contributions into a statistically and computationally efficient solution to our problem.
- April 30, 2024 @ Bâtiment IMAG (442) -- Seminar Rémi Castera
Correlation of Rankings in Matching Markets