-
Notifications
You must be signed in to change notification settings - Fork 153
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
some implementation issues #92
Comments
Hi, thanks for the feedback!
|
Hello! @eleurent Thank you so much for your reply, but I'm still a little confused.
I'm sorry if my statement is not clear, because I haven't done this kind of work before, and my thinking is a little confused.. so I come here again to ask you for advice. Really looking forward to your reply, this is very helpful for me, thank you! |
|
@eleurent |
No worries at all, glad you found this helpful! |
Hello @eleurent ! I would like to ask some questions, so I reopened this issue. |
Hi @eleurent! Sorry I have another question.
Look forward to your answer! Thank you!! |
They do! It's just that the we don't know the mapping between vehicle i in the attention matrix and the vehicles in the scene, since the vehicles ids are not provided in the observation. So in order to draw the attention edges, I'm mapping rows of the observation to vehicles in the scene based on their x,y coordinate features.
Yes we typically limit the observation to the nearest few cars, e.g. 15 vehicles. We observe empirically that the attention is useful by enabling the model to focus its computations on the 1-2 most relevant vehicles at any given time, which leads to better decisions. It is also invariant to permutations of the vehicles ordering, unlike other architectures such as MLPs. |
Sorry for the late reply! Thank you for answering my confusion! |
Hi! @eleurent |
I think there can be two roles:
These are just hypotheses of course, the function that the attention layer ends up implementing is emerging through learning. You could very well do the experiment of removing the ego-vehicle from the available tokens, and see if this degrades performance. (you'd probably need to keep the residual connection though, we still want the ego-vehicle's feature to be available for the final decision. |
Got it! Thank you!! |
Hi! First of all thank you for sharing such a great project! This gave me a lot of inspiration! Really appreciate!
I have some questions I would like to ask you.
I read your paper "Social Attention for Autonomous Decision-Making in Dense Traffic", and you mainly proposed an "attention-based neural network architecture". But in this repo, what is the purpose of implementing so many agents (MCTS, DQN...)? Different ways to implement this ‘attention architecture’?
Where is the scripts/analyze.py file? Has it been superseded?
As an RL rookie, I would like to ask if the ‘attention architecture’ you proposed can be used in other RL algorithms? As an example, suppose I have trained an RL algorithm called 'ABC' to control all the autonomous vehicles in the scene.
Now I want to add your proposed 'attention architecture' to it, so that each vehicle knows which vehicles around it should pay the most attention. Finally the 'ABC' algorithm is used to train the whole model.
I want to know is this possible? How should I integrate 'attention architecture' into 'ABC'?
Looking forward to your reply! Thanks!
The text was updated successfully, but these errors were encountered: