Home

Regularization Methods for Deep Learning


Speaker

Haizhao Yang, National University of Singapore

Time

2019-06-06 10:30:00 ~ 2019-06-06 11:30:00

Location

Room 3-412, SEIEE Building

Host

Jingwen Leng, Assistant Professor, John Hopcroft Center for Computer Science

Abstract

Overfitting frequently occurs in deep learning. In this talk, we introduce two novel regularization methods: Drop-Activation and DIA-Net, to reduce overfitting and improve generalization. The key idea of Drop-Activation is to weaken the nonlinear effect of ReLU activation functions by setting them to be identities randomly during training time; while the crucial idea of the DIA-Net is to enhance the communication of network attention via recurrent neural networks. Experimental results on CIFAR-10, CIFAR-100, SVHN, EMNIST, and ImageNet show that these regularization methods generally improve the performance of popular neural network architectures, and are harmonic with other regularization algorithms.

Bio

Dr. Haizhao Yang is currently an assistant professor in the Department of Mathematics and an affiliate of the Institute of Data Science at the National University of Singapore (NUS). Prior to joining NUS, Dr. Yang was a visiting assistant professor at Duke University from 2015 to 2017. Dr. Yang obtained his Ph.D. at Stanford University in 2015, M.S. at the University of Texas at Austin in 2012, and B.S. at Shanghai Jiao Tong University in 2010 all in applied mathematics.

© John Hopcroft Center for Computer Science, Shanghai Jiao Tong University
分享到

地址:上海市东川路800号上海交通大学软件大楼专家楼
邮箱:jhc@sjtu.edu.cn 电话:021-54740299
邮编:200240