论文标题
有效的最大编码率通过变异形式降低
Efficient Maximal Coding Rate Reduction by Variational Forms
论文作者
论文摘要
最近提出了最大编码率降低(MCR $^2 $)的原理,作为学习判别性低维度结构到固有到高维数据的训练目标,以便比标准方法更强大的训练,例如交叉透镜最小化。但是,尽管对MCR $^2 $培训显示出优势,但由于需要评估和区分大量的日志确定术语,而MCR $^2 $遭受了巨大的计算成本,这些术语随着类的数量线性增长。通过利用矩阵的各种频谱函数的各种形式,我们将MCR $^2 $的客观重新定位为可以大大扩展的形式,而不会损害训练精度。图像分类的实验表明,我们提出的配方会导致高于优化原始MCR $^2 $客观的显着速度,并且通常会导致更高质量的学习表示。此外,我们的方法可能具有独立的兴趣,这些模型需要计算对数确定形式(例如系统识别或正常化流量模型)。
The principle of Maximal Coding Rate Reduction (MCR$^2$) has recently been proposed as a training objective for learning discriminative low-dimensional structures intrinsic to high-dimensional data to allow for more robust training than standard approaches, such as cross-entropy minimization. However, despite the advantages that have been shown for MCR$^2$ training, MCR$^2$ suffers from a significant computational cost due to the need to evaluate and differentiate a significant number of log-determinant terms that grows linearly with the number of classes. By taking advantage of variational forms of spectral functions of a matrix, we reformulate the MCR$^2$ objective to a form that can scale significantly without compromising training accuracy. Experiments in image classification demonstrate that our proposed formulation results in a significant speed up over optimizing the original MCR$^2$ objective directly and often results in higher quality learned representations. Further, our approach may be of independent interest in other models that require computation of log-determinant forms, such as in system identification or normalizing flow models.