-
Notifications
You must be signed in to change notification settings - Fork 45
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
No Crawler information on running crawler #24
Comments
you can log into the container then run the cron rake task
or by docker compose:
|
how can one know when the crawl has completed? |
I followed the instruction on Readme file, so my database already has the data from the January pull (i.e. the count() would return values). It appears that the database is not being refreshed with the latest data, hence why I'm not sure the crawl is actually active. FYI, I am also new to Rails and Docker. |
I didn't pre populate the data as specified in the README file but you should be able to run the crawler in any case. He called the task in this case it was triggered by the linux os in the docker containter. see: Its overkill for everybody who clones the repo to do this on a daily basis so just run it manually once in a while: You will notice if its running as there is terminal output and its very intensive on your machine If you look in lib/tasks/cron.rake you will see:
|
I'm guessing the Crawler is run automatically when you execute the command "docker-compose up"? I tried the command "docker-compose exec app rake cron" and get rake aborted! Tasks: TOP => cron (See full trace by running task with --trace) |
I'm getting that as well trying to run it a second time. I think its got something to do with Crawler state. Give me a minute... |
run this in rails console app/models/crawl.rb is_active checks for these states and will exit with run this in rails console then run the cron task: |
The second command generated some error messages - not sure if it's normal. Then running the cron task showed the same "Crawl is already running" message. By the way, really appreciate you helping out! Below is the output from executing the commands in rails. Loading development environment (Rails 5.2.2) from /usr/local/bundle/gems/activerecord-5.2.2/lib/active_record/connection_adapters/postgresql_adapter.rb:611:in `async_exec_params' |
try Crawl.find(2810).destroy use any id returned by Crawl.where(state: [:init, :running, :paused]) or try reinstalling everything without importing the old data... |
Thanks for the suggestion, I'm not sure why it didn't work. I finally just deleted the db image and it's crawling finally! yay! thanks again for all your help. |
Where did you find the db image? I'm having the same issue with my crawler @chimemeh |
i believe it will be created on initialization of the rails app or on the first crawl. What do you have so far? |
there is only a brief mention of the crawler but no instructions on how to run the crawler. if you could post the commands to run the crawler id be more then happy to update the read me with the information and a guide on how to use it.
The text was updated successfully, but these errors were encountered: