Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] Auto-regressive manner policy network? #2010

Closed
4 tasks done
wadmes opened this issue Sep 17, 2024 · 1 comment
Closed
4 tasks done

[Question] Auto-regressive manner policy network? #2010

wadmes opened this issue Sep 17, 2024 · 1 comment
Labels
question Further information is requested

Comments

@wadmes
Copy link

wadmes commented Sep 17, 2024

❓ Question

Hi there,

I am trying to implement a autoregressive policy network, for example,
in each action, I need to sample target object, and then use the target object embedding to sample the action type, and finally output the action parameters.

But it looks it is impossible to achive this since what _build_mlp_extractor in MaskableActorCriticPolicy asks it to return a action embedding and value embedding.

Is there a good way to achieve the feature?

Example figure:

image

Thanks!

Checklist

@wadmes wadmes added the question Further information is requested label Sep 17, 2024
@araffin
Copy link
Member

araffin commented Sep 18, 2024

Is there a good way to achieve the feature?

For something custom like that, it's probably better to fork SB3 / use CleanRL as a starting point.

@araffin araffin closed this as completed Oct 7, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants