INTERPRETABLE CLASSIFICATION VIA SUPERVISED VARIATIONAL ...
openreview.net › pdfconnect to a variational autoencoder (VAE) to learn an embedding of the data that the tree can classify with low expected loss. The expected loss of the DDT is differentiable, so standard gradient-based methods may be applied in training. Since we work in a supervised learning setting, it is natural to exploit the label information when
[1603.02514] Variational Autoencoders for Semi-supervised ...
https://arxiv.org/abs/1603.0251408.03.2016 · Although semi-supervised variational autoencoder (SemiVAE) works in image classification task, it fails in text classification task if using vanilla LSTM as its decoder. From a perspective of reinforcement learning, it is verified that the decoder's capability to distinguish between different categorical labels is essential. Therefore, Semi-supervised Sequential …
[1902.00220] A Classification Supervised Auto-Encoder Based ...
arxiv.org › abs › 1902Feb 01, 2019 · Classic variational autoencoders are used to learn complex data distributions, that are built on standard function approximators. Especially, VAE has shown promise on a lot of complex task. In this paper, a new autoencoder model - classification supervised autoencoder (CSAE) based on predefined evenly-distributed class centroids (PEDCC) is proposed. Our method uses PEDCC of latent variables to ...
Self-Supervised Variational Auto-Encoders | OpenReview
openreview.net › forumSep 28, 2020 · Variational Auto-Encoders (VAEs) constitute a single framework to achieve these goals. Here, we present a novel class of generative models, called self-supervised Variational Auto-Encoder (selfVAE), that utilizes deterministic and discrete transformations of data. This class of models allows performing both conditional and unconditional ...
INTERPRETABLE CLASSIFICATION VIA SUPERVISED …
https://openreview.net/pdf?id=rJhR_pxCZconnect to a variational autoencoder (VAE) to learn an embedding of the data that the tree can classify with low expected loss. The expected loss of the DDT is differentiable, so standard gradient-based methods may be applied in training. Since we work in a supervised learning setting, it is natural to exploit the label information when