百家乐开户-百家乐官网开户-机械百家乐技巧

網(wǎng)站頁(yè)面已加載完成

由于您當(dāng)前的瀏覽器版本過(guò)低,存在安全隱患。建議您盡快更新,以便獲取更好的體驗(yàn)。推薦使用最新版Chrome、Firefox、Opera、Edge

Chrome

Firefox

Opera

Edge

ENG

當(dāng)前位置: 首頁(yè) · 學(xué)術(shù)交流 · 正文

學(xué)術(shù)交流

【學(xué)術(shù)報(bào)告】研究生“靈犀學(xué)術(shù)殿堂”第560期之徐揚(yáng)揚(yáng)教授報(bào)告會(huì)通知

發(fā)布時(shí)間:2020年09月18日 來(lái)源:研究生院 點(diǎn)擊數(shù):

全校師生:

我校定于2020年09月26日舉辦研究生靈犀學(xué)術(shù)殿堂——徐揚(yáng)揚(yáng)教授報(bào)告會(huì),現(xiàn)將有關(guān)事項(xiàng)通知如下:

1.報(bào)告會(huì)簡(jiǎn)介

報(bào)告人:徐揚(yáng)揚(yáng)教授

時(shí)間:2020年09月26日(星期六)10:00

地點(diǎn):騰訊會(huì)議(會(huì)議號(hào):688 912 696)

報(bào)告題目:Accelerating stochastic gradient methods

內(nèi)容簡(jiǎn)介:Stochastic gradient method has been extensively used to train machine learning models, in particular for deep learning. Various techniques have been applied to accelerate stochastic gradient methods, either numerically or theoretically, such as momentum acceleration and adapting learning rates. In this talk, I will present two ways to accelerate stochastic gradient methods. The first one is to accelerate the popular adaptive (Adam-type) stochastic gradient method by asynchronous (async) parallel computing. Numerically, async-parallel computing can have significantly higher parallelization speed-up than its sync-parallel counterpart. Several previous works have studied async-parallel non-adaptive stochastic gradient methods. However, a non-adaptive stochastic gradient method often converges significantly slower than an adaptive one. I will show that our async-parallel adaptive stochastic gradient method can have near-linear speed-up on top of the fast convergence of an adaptive stochastic gradient method. In the second part, I will present a momentum-accelerated proximal stochastic gradient method. It can have provably faster convergence than a standard proximal stochastic gradient method. I will also show experimental results to demonstrate its superiority on training a sparse deep learning model.

2.歡迎各學(xué)院師生前來(lái)聽(tīng)報(bào)告。報(bào)告會(huì)期間請(qǐng)關(guān)閉手機(jī)或?qū)⑹謾C(jī)調(diào)至靜音模式。

西北工業(yè)大學(xué)黨委學(xué)生工作部

數(shù)學(xué)與統(tǒng)計(jì)學(xué)院

復(fù)雜系統(tǒng)動(dòng)力學(xué)與控制工信部重點(diǎn)實(shí)驗(yàn)室

2020年9月18日

報(bào)告人簡(jiǎn)介

Dr. Yangyang Xu(徐揚(yáng)揚(yáng)) is now a tenure-track assistant professor in the Department of Mathematical Sciences at Rensselaer Polytechnic Institute. He received his B.S. in Computational Mathematics from Nanjing University in 2007, M.S. in Operations Research from Chinese Academy of Sciences in 2010, and Ph.D from the Department of Computational and Applied Mathematics at Rice University in 2014. His research interests are optimization theory and methods and their applications such as in machine learning, statistics, and signal processing. He developed optimization algorithms for compressed sensing, matrix completion, and tensor factorization and learning. Recently, his research focuses on first-order methods, operator splitting, stochastic optimization methods, and high performance parallel computing. He has published over 30 papers in prestigious journals and conference proceedings. He was awarded the gold medal in 2017 International Consortium of Chinese Mathematicians.