You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is the current implementation of mobility regularity-aware loss correct? Now both $L_d$ and $L_p$ are directly added to the loss for calculating policy gradient, but they won't produce any gradient on the input (generated sequences, since they are discrete values).
I guess the correct way is to add them to the reward instead. Is that right?
The text was updated successfully, but these errors were encountered:
MoveSim/code/main.py
Line 165 in 93e6837
Is the current implementation of mobility regularity-aware loss correct? Now both$L_d$ and $L_p$ are directly added to the loss for calculating policy gradient, but they won't produce any gradient on the input (generated sequences, since they are discrete values).
I guess the correct way is to add them to the reward instead. Is that right?
The text was updated successfully, but these errors were encountered: