The double descent behavior In two layer neural network for binary classification

时间:2025-10-30浏览:10设置

讲座题目:The double descent behavior In two layer neural network for binary classification

主讲人:桑海林 教授

主持人:徐方军 教授

开始时间:2025-11-10:00

讲座地址:腾讯会议:511-511-721;  密码: 0906

主办单位:统计学院


报告人简介:

      桑海林,美国密西西比大学数学系教授,2008年博士毕业于美国康涅狄格大学。 主要从事随机场、非参统计、经验过程,机器学习等方面研究。在Stochastic Processes and Their Applications, Statistica Sinica, Journal of Time Series Analysis, Journal of Nonparametric Statistics等一流学术期刊发表论文近30篇。


报告内容:

Recent studies observed a surprising concept about test error called  the double descent phenomenon where the increasing model complexity de creases the test error rst and then the error increases and decreases again.  To observe this, we worked on a two-layer neural network model with a  ReLU activation function designed for binary classi cation under supervised  learning. Our aim was to observe and nd the mathematical theory behind  the double descent behavior of the test error in the model for varying over parameterization and under-parameterization ratios. We have been able to  derive a closed-form solution for the test error of the model and a theorem  to nd the parameters with optimal empirical loss when model complexity  increases. We proved the existence of the double descent phenomenon in  our model for square loss function using the theorems derived.  This talk is based on the work jointly with Chathurika Abeykoon and  Aleksandr Beknazaryan.

返回原图
/