-
Notifications
You must be signed in to change notification settings - Fork 856
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Dr-Elephant not fetching RUNNING spark application (only succeeded and failed applications are fetched) #696
Comments
Hi I would like to know why the dr-elephant does not support fetching RUNNING applications natively. Is there a reason for this choice (performance, technical constraints, ...). Thx |
Initially, Dr.Elephant was designed to profile a Hadoop job after finishes. This idea stayed with the Spark Heuristics too. But with the increased demand Spark streaming we do know the importance of a tool to track your jobs' performance. The reason for not supporting the Spark Streaming applications is the large logs. Currently, SHS doesn't provide any incremental parsing of logs, so if Dr.Elephant analyzes a RUNNING application at some short interval then it has to parse the whole logs every time and with Streaming jobs, this issue becomes critical as their log size keeps on increasing. This will hog the Dr.Elephant's resources and lead to delays in report generation etc. With the batch jobs, the need for real-time profiling is not that missed, so there are challenges to support RUNNING apps in Dr.Elephant. I would be glad to know how you are approaching these challenges and would try to provide any needed assistance from my end. |
Hi, We try to deal with these problems in the following way :
|
Hi @nelhaj Could you share how you modified Dr. Elephant to fetch and analyze running applications? Any additional tips or considerations for implementing this feature. Your insights would be greatly appreciated. Thank you |
Hi,
Dr-Elephant is only fetching completed applications (filtered by SUCCEEDED or FAILED status).
Our spark streaming applications are always RUNNING non-stop (except for weekly restarts).
We want to be able to analyze them and generate real time heuristics.
Why does dr-elephant exclude running application ?
Is there a way to include them when fetching jobs list?
More details:
We are using SparkFetcher.
Dr. Elephant gets list of only succeeded and failed applications from Yarn History Server API:
Running spark application are available in both YARN HS and Spark HS. I can retrieve log events by accessing http://{SHS_HOST}/api/v1/applications/application_xxxxxxxxxx_xxxxxx/1/logs
Thank you
The text was updated successfully, but these errors were encountered: