ydataai / ydata-synthetic

Synthetic data generators for tabular and time-series data
https://docs.synthetic.ydata.ai
MIT License
1.41k stars 234 forks source link

Latent Space Auto Encoder #165

Open Franz0808 opened 2 years ago

Franz0808 commented 2 years ago

Hello,

I am wondering about the autoencoder, especially the embedder that transforms the data into the latent space. If I understood it correctly, the embedders task is to compress the dimensionality of the original data into a lower dimension in order to learn the underlying structure.

But lets say we have the stock data (6 columns), a rolling windows sice of 24 and a batch sice of 50. We give the embedder data of the shape (50,24,6). In the code we use 24 neurons per layer... so that the output of the embedder is 50,24,24....

Wouldn't that imply that instead of compressing the data to a lower dimensional space, we kind of extend the latent space? Dont we need a bottleneck layer in th autoencoder of less than 6 in order to achieve the dimensionaliy reduction we are looking for?

Thanks in advance for your response. Kind regards, Franz

fabclmnt commented 1 year ago

Hi Franz,

have you checked the information and details from the author regarding the architecture decision?

https://proceedings.neurips.cc/paper/2019/file/c9efe5f26cd17ba6216bbe2a7d26d490-Paper.pdf

We are happy to discuss more details on the architecture in our Discord Community (https://discord.com/invite/mw7xjJ7b7s)