You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I find that the method of fusing the features is not the same as in the paper that relies on the attention mechanism score to select the features of the corresponding agent to fuse .
In the code, during training, the fused features are obtained directly using the attention mechanism, and it seems that the returned agent number, i.e., action_argmax , is not used subsequently.
I find that the method of fusing the features is not the same as in the paper that relies on the attention mechanism score to select the features of the corresponding agent to fuse .
In the code, during training, the fused features are obtained directly using the attention mechanism, and it seems that the returned agent number, i.e., action_argmax , is not used subsequently.
MultiAgentPerception/ptsemseg/trainer.py
Line 391 in 4ef3005
And you can see that the use of torch.argmax during training does not match the thesis
MultiAgentPerception/ptsemseg/models/agent.py
Lines 625 to 627 in 4ef3005
https://github.com/GT-RIPL/MultiAgentPerception/blob/4ef300547a7f7af2676a034f7cf742b009f57d99/ptsemseg/trainer.py#L384C1-L398C1
And instead of using commun_label during training, it was used as a metric during validation and testing.
I feel that the log_action, action_argmax are really meaningless, and the features obtained during training are not all the features of a vehicle, but are obtained by using the attention method. Instead of selecting all the features of a certain car, in the test, new features are added after the attention is used.
https://github.com/GT-RIPL/MultiAgentPerception/blob/4ef300547a7f7af2676a034f7cf742b009f57d99/ptsemseg/models/agent.py#L629C1-L651C56
The text was updated successfully, but these errors were encountered: