论文标题
清醒地看一下无监督的分解表示的学习及其评估
A Sober Look at the Unsupervised Learning of Disentangled Representations and their Evaluation
论文作者
论文摘要
\ emph {无监督的}学习\ emph {distangled}表示背后的想法是,现实世界数据是由一些解释性因素生成的变化因素,这些因素可以通过无处不在的学习算法来恢复。在本文中,我们清醒地研究了该领域的最新进展,并挑战了一些共同的假设。我们首先从理论上表明,如果没有模型和数据上的归纳偏见,对分解表示的无监督学习是不可能的。然后,我们在对八个数据集的可重复的大规模实验研究中培训了涵盖最突出方法和评估指标的$ 14000 $模型。我们观察到,尽管不同的方法成功地强制执行相应损失的“鼓励”属性,但在没有监督的情况下似乎无法识别出良好的模型。此外,不同的评估指标并不总是就应该被认为是“分解”的内容并在估计中表现出系统的差异。最后,增加的分解似乎并不一定会导致下游任务的学习样本复杂性降低。我们的结果表明,未来关于解开学习的工作应明确有关归纳偏见和(隐性)监督的作用,调查实施学习表示表示的具体益处,并考虑涵盖几个数据集的可重复实验设置。
The idea behind the \emph{unsupervised} learning of \emph{disentangled} representations is that real-world data is generated by a few explanatory factors of variation which can be recovered by unsupervised learning algorithms. In this paper, we provide a sober look at recent progress in the field and challenge some common assumptions. We first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without inductive biases on both the models and the data. Then, we train over $14000$ models covering most prominent methods and evaluation metrics in a reproducible large-scale experimental study on eight data sets. We observe that while the different methods successfully enforce properties "encouraged" by the corresponding losses, well-disentangled models seemingly cannot be identified without supervision. Furthermore, different evaluation metrics do not always agree on what should be considered "disentangled" and exhibit systematic differences in the estimation. Finally, increased disentanglement does not seem to necessarily lead to a decreased sample complexity of learning for downstream tasks. Our results suggest that future work on disentanglement learning should be explicit about the role of inductive biases and (implicit) supervision, investigate concrete benefits of enforcing disentanglement of the learned representations, and consider a reproducible experimental setup covering several data sets.