About this Event
As the continuous limit of many discretized algorithms, PDEs can provide a qualitative description of algorithm's behavior and give principled theoretical insight into many mysteries in machine learning. In this talk, I will give a theoretical interpretation of several machine learning algorithms using Fokker-Planck (FP) equations. In the first one, we provide a mathematically rigorous explanation of why resampling outperforms reweighting in correcting biased data when stochastic gradient-type algorithms are used in training. In the second one, we propose a new method to alleviate the double sampling problem in model-free reinforcement learning, where the FP equation is used to do error analysis for the algorithm. In the last one, inspired by an interactive particle system whose mean-field limit is a non-linear FP equation, we develop an efficient gradient-free method that finds the global minimum exponentially fast.
Event Details
Dial-In Information
Join by telephone (use any number to dial in)
+1 929 436 2866
+1 301 715 8592
+1 312 626 6799
+1 669 900 6833
+1 253 215 8782
+1 346 248 7799
International numbers available: https://harvard.zoom.us/u/acvB7Y7Eyx
One tap mobile: +19294362866,,96063768850# US (New York)
Join by SIP conference room system
Meeting ID: 960 6376 8850