Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Make simulate_pixels.py more robust against huge tracks datasets #102

Conversation

peter-madigan
Copy link
Member

@peter-madigan peter-madigan commented Oct 6, 2022

  • Filters tracks dataset against declared active TPCs
  • Restructures simulate_pixels.py to create a minimal memory footprint prior to filtering tracks dataset
  • Add segment_id to tracks dataset in order to preserve MC truth backtracking

For the 2x2 file that @krwood provided me, I no longer encounter memory overruns on a single CORI 16GB Tesla V100 gpu with a tracks dataset containing 20M entries.

The tracks dataset is still loaded up front which takes approx ~30s. The dataset is then filtered on the CPU (~20s), which could probably be sped up by copying the xyz data to the GPU. The subsequent simulation then only takes about another 20s.

@peter-madigan peter-madigan requested a review from krwood October 6, 2022 01:38
@krwood
Copy link
Member

krwood commented Oct 6, 2022

I see the same improved performance.

@krwood krwood merged commit 23973b3 into DUNE:master Oct 6, 2022
@peter-madigan peter-madigan deleted the fix/improve-stability-for-large-tracks-datasets branch October 6, 2022 21:21
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants