-
-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Action fails, but no error? Appears incomplete - no summary. #104
Comments
I am not sure what is the issue exactly but you seem to have this issue for a while. The way I would debug this is to test the same workflow using the Python module locally and see if that works. If so then maybe I would lower the number of workers, or set it to 1 and see if that solves the issue. |
Thanks so much for the quick response and suggestions. Admittedly, we're not sure how to run actions locally (we're a group of biologists, so not well-versed in software development stuff), but we'll poke around the web and report back. |
Hey @kubu4 ! You shouldn't need to poke around the web - urlchecker is a command line python tool, and there are instructions for install and usage here: https://github.com/urlstechie/urlchecker-python The action is simply running that under the hood. Let us know if you have any questions! I work with a lot of biologists. :) |
Ha! Thanks! |
Brief update: I think it's a memory issue. I ran I didn't let it finish because some other people were trying to use the computer for some other tasks and I, essentially, locked it up. Possible solution? Reducing the number of workers, per @SuperKogito's suggestion? |
That seems strange - how many files are you checking (and what is a qmd extension)? Try adding |
Thousands of files. Many links are to large files (multi-GB in size). Would that have an impact on how this action runs? |
Yes likely - maybe try out testing a smaller subset of the files first and see at what size it starts to not work? |
this is actually consistent with the memory overflow; the files are loaded and scanned for urls. This definitely will require a lot of RAM if your files are too big/ too many. Using multiple workers will only make this worse, hence I mentioned using one worker. Using '--serial' per @vsoch recommendation is also a possible solution but if your files are too big, it will be hard to escape this. Especially, if the memory is not flushed as soon as the links are extracted. |
Likely we need a fix that processes them in batches (and doesn't try to load everything into memory at once). |
You could also just target runs on separate subdirectories (one at a time or in a matrix), depending on how large your repository is. |
I suspect that this has something to do with the memory management and garbage collection.
@vsoch generators could be a good fix here, what do you think of this? |
@SuperKogito my first suggestion to @kubu4 is to try processing in batches (e.g., multiple runs on different roots, and that can be put into an action matrix). If that doesn't work, then I think we should add some kind of support to handle that internally. |
Whenever the
urlchecker-action
runs, it fails. The end of the log file appears to be incomplete, as it doesn't provide any summary and/or error messages.This is what my workflow file looks like:
I'm not sure how to troubleshoot. Are there log files that get generated somewhere that I can look through?
The text was updated successfully, but these errors were encountered: