Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

No support for NVILA? #839

Open
1 task done
George-Polya opened this issue Feb 10, 2025 · 2 comments
Open
1 task done

No support for NVILA? #839

George-Polya opened this issue Feb 10, 2025 · 2 comments
Labels
question Further information is requested

Comments

@George-Polya
Copy link

George-Polya commented Feb 10, 2025

Search before asking

  • I have searched the jetson-containers issues and found no similar feature requests.

Question

I used VLM of Jetson Platform Service. In here, I saw your mlc-llm patch.

I tried to use Efficient-Large-Model/NVILA-8B.

Then, I saw the error below.

Traceback (most recent call last):
  File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
    return _run_code(code, main_globals, None,
  File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
    exec(code, run_globals)
  File "/usr/local/lib/python3.10/dist-packages/mlc_llm/build.py", line 47, in <module>
    main()
  File "/usr/local/lib/python3.10/dist-packages/mlc_llm/build.py", line 41, in main
    parsed_args = core._parse_args(parsed_args)  # pylint: disable=protected-access
  File "/usr/local/lib/python3.10/dist-packages/mlc_llm/core.py", line 444, in _parse_args
    parsed = _setup_model_path(parsed)
  File "/usr/local/lib/python3.10/dist-packages/mlc_llm/core.py", line 494, in _setup_model_path
    validate_config(args.model_path)
  File "/usr/local/lib/python3.10/dist-packages/mlc_llm/core.py", line 538, in validate_config
    config["model_type"] in utils.supported_model_types
AssertionError: Model type qwen2 not supported.

Maybe, I think that's because mlc patches. How can I use NVILA?

Additional

No response

@George-Polya George-Polya added the question Further information is requested label Feb 10, 2025
@dusty-nv
Copy link
Owner

Hi @George-Polya , I rebuilt it last week in AWQ: 75c1ab1 dustynv/awq:r36.4.0

the NVILA repo supports it in HF and AWQ TinyChat. I have the action to profile it and add TinyChat support to OpenAI server. When it is more ready I will retag dustynv/awq as dustynv/vila since those are circular dependencies now.

@George-Polya
Copy link
Author

George-Polya commented Feb 10, 2025

Thank you for your response. I will wait for it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants