Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Posterior computation #2

Open
sai-prasanna opened this issue May 24, 2024 · 1 comment
Open

Posterior computation #2

sai-prasanna opened this issue May 24, 2024 · 1 comment

Comments

@sai-prasanna
Copy link

In your posterior, you use the stochastic state of the prior. But in RSSM they only use the deterministic state, and observation embedding. Since the prior's stochastic state is just a function of the deterministic state, it won't have extra information to condition upon. And using the stochastic state sample might hurt computing the posterior because of the sampling noise.

I am checking in case there is some other deeper reason to use it.

x = torch.cat([prior_state['stoch'], prior_state['deter'], obs_embed], dim=-1)

@mazpie
Copy link
Owner

mazpie commented Jun 6, 2024

Hi @sai-prasanna,
you're right that there is a subtle difference with the original RSSM.

However, I would not expect any major differences as the information to condition upon is contained in the deterministic state, as you pointed out.

The stochastic state might either be helpful (it is a more noisy estimate of the state) or be ignored by the network, if it doesn't contain any useful information (e.g. if you just concatenate random noise to the inputs of a network, the network quickly learns to ignore it)

I hope this answers your question!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants