-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
WIP: Add support for NVIDIA nvc++ compiler #17000
base: master
Are you sure you want to change the base?
Conversation
Test Results 18 files 18 suites 4d 15h 23m 21s ⏱️ Results for commit f41386a. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Have you tried to run the test suite for ROOT with this pull request and nvcc? My worry is that we have to execute cuda code in the interpreter through clang and some changes in the interpreter's cmake files seem to go towards getting the nvcc runtime somehow in cling.
@@ -9,7 +9,14 @@ | |||
#--------------------------------------------------------------------------------------------------- | |||
|
|||
if(NOT CMAKE_CXX_COMPILER_ID MATCHES "(Apple|)Clang|GNU|Intel|MSVC") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Don't we need a check here for NVHPC
too?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
As far as I understand this condition filters out supported compilers. As NVHPC
is only partially supported, it must enter this condition - note the NOT
in front.
@cgleggett so this means the most recent |
what is the recommended way to test a build? I was able to do I/O and read/write root files, but that obviously doesn't cover everything. |
If you would like to run the full test suite, you can configure with:
and run You can also start without the |
building the tests have revealed a few more issues with support for flags in nvc++. WIPing until these are addressed. |
@pcanal : running the tests showed 413 failures with nvc++. However, making a build on identical source with gcc 13.3 and then running the tests showed 181 failures. Surprisingly, the overlap was only 15 tests which failed in BOTH nvc++ and gcc. So, any suggestions on which tests are critical and I should really look at? |
@cgleggett I assume you have the same branch/corresponding commits from the root and roottest repository. If so all tests are meant to succeed with gcc. A first steps to ensure that all the test from |
@pcanal: I'm building the same sources for both compilers, using should this produce zero failed tests with gcc13? I was trying to build a smaller library without a lot of the graphics bits. |
It definitively should ... albeit the options |
I forked from master at e344b22. the commits in the fork should have no impact on the gcc builds. Should I have forked from a different branch to test, eg Building with just |
In my case, I see the following failures:
The 682 is due to a missing protection against building without OpenGL. The 270 might be because of the missing VDT. The later one is due (in my case) to the missing Davix ... but seems to work on other node without davix. Many python test will fails if you are missing the pip install of |
Can you give me the list? Actually, just send me the full
You need to ensure that you have a corresponding roottest. Do:
|
Picking this up after the break.... I've installed all the requirements in Attached is the list of failed tests, as well as LastTest.log and LastTestsFailed.log for gcc13.3.0. I'll attach the ones for nvc++ in another post. |
for gcc 13.3.0: |
There is something very wrong with the python interactions, a lot of the test seems to crash badly:
|
here's the output of valgrind:
running with gdb doesn't show much:
the nvc++ build runs this test fine |
Hack :( ... okay, let's side steps this for a round and see what else might be failing. |
probably unrelated, but I get this error message when building:
the compile line looks like:
there are in fact another one:
if I remove the now to try the ctest.... |
The tests that failed with
|
Is fixed by #17459 A workaround is to reconfigure ( |
is minor:
@couet What can be done here? |
Are either a missing install or mismatch in version:
@guitargeek What can be done here? |
Are more 'interesting' and may or may not be due to how Tensor flow was built.
However it is not clear what the real error is. It ends with:
|
Probably due to the python/pyroot issue. Let's ignore those for now. |
The other test fails because of the prior failures. |
The conclusion is that in addition to the issue with python/pyroot, the other issues are Keras or TensorFlow related. Were the results #17000 (comment) with nvcc, gcc or both? |
Did you install the test environment with ROOTs requirements.txt file? Note that there are maximum supported torch and TensorFlow versions. |
I installed python 3.11 and then rebuilt (without the So maybe there's an issue with the default python 3.9 in Alma9? |
This Pull request:
Adds support for NVIDIA's nvc++ compiler
Fixes #16975
Changes or fixes:
This allows clean compilation on ARM (NVIDIA Grace) CPUs.