Skip to content
This repository has been archived by the owner on May 24, 2020. It is now read-only.

Model Latent Layer Dimensionaliy? #6

Open
Hamcastle opened this issue Jan 12, 2019 · 1 comment
Open

Model Latent Layer Dimensionaliy? #6

Hamcastle opened this issue Jan 12, 2019 · 1 comment

Comments

@Hamcastle
Copy link

Hi,

The default config for the vae has an intermediate layer size of 32 and a latent layer size of 100. The example data are 3*13, so the input dimension is 39. Is there some reason that unlike most autoencoders, the "bottleneck" middle layer is larger than the input? There are auto-encoder architectures that do this, but usually they require weight regularization somewhere in the layer sequence.

@twairball
Copy link
Owner

no real reason - it just follows defaults from the keras VAE example.

The bundled example is probably not a good example as per your point...

@twairball twairball reopened this Jan 14, 2019
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants