说明:We study a new aggregation operator for gradients coming from a mini-batch for stochastic
gradient (SG) methods that allows a significant speed-up in the case of sparse optimization
problems. We call this method AdaBatch and it only requires a few <sinat_20409085> 上传 | 大小:393kb