-
Notifications
You must be signed in to change notification settings - Fork 204
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The Total waiting time of my system is constantly increasing!! (in the csv, I saved with the method save_csv()) #173
Comments
could u give me your email ? We can discuss this issue together |
|
My Email: [email protected] |
Hi, by any chance, are you perhaps introducing too many random trips? It is possible that you created a simulation in which it is impossible to serve all the vehicles, and they keep being queued to enter in the network. |
I thought of this and I started increasing the intervals between spawning of each vehicles: |
hey @LucasAlegre thanks, I am trying some variation by reducing the randomness even more like 90 or 45 vehicles in the environment according to your suggestion. I am seeing some good results in terms of increased average speed. And I am also trying some various reward functions too. I will update you once I finish my trials. Thanks |
Hey man, I have tried these variations:
while generating the randomTrips to reduce the no of vehicles at a time. But they result in very less no of no of vehicles in the environment therefore leaving empty intersections or single intersection with stagnant queues. So is there any way to reduce this randomness (as you mentioned), and rectify this flaw from my approach? |
Hii.. Finally I solved the issue after training the SAC in multi-agent environment for almost 1000 episodes. I think the reduced randomness (-p 20) and waiting-time reward function and the 8 hour training process in my 4GB GPU did it. As you can see, I am getting some decent trends in my first graph. And my rewards have also been improved overall.. (abscissa - no of episodes, ordinate - rewards)
Thank you guys very much for your suggestions!! |
I'm glad that you were able to run optimize the network successfully! :) |
I have used the DQN, A2C, PPO from stablebaselines3 package, along with this sumo-rl package. I am using it with my own sumo network, with 4 traffic signal intersections. I have also generated random trips for my net file using randomTrips.py. This is the command I used to generate random trips.
python $SUMO_HOME/tools/randomTrips.py -n nets/2x2/2x2.net.xml -r nets/2x2/2x2-p20.rou.xml -e 3600 -p 20
this is my net file (2x2 from the nets folder):
this is the plot results i got with the csv saved:
for single intersection the results are looking okay.
But as soon as I use a multi-agent environment. This occurs. I have also used petting zoo environment provided in this package along with my custom model, I have set my single agent parameter as false. But still the results remain the same.
These are the results of my custom SAC Multi Agent model.
I have also tried various reward functions like differential waiting time, average speed, queue, pressure but no change.
I have no idea, what to do.. I have been stuck on this for almost 2 months.
Please, guide me. Thanks a lot.
The text was updated successfully, but these errors were encountered: