Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about integration with DeepSpeed-Ulysses #679

Closed
zigzagcai opened this issue Nov 15, 2024 · 2 comments
Closed

Question about integration with DeepSpeed-Ulysses #679

zigzagcai opened this issue Nov 15, 2024 · 2 comments
Labels
question Further information is requested

Comments

@zigzagcai
Copy link

Hi developers,

Thanks for such a great project that can demonstrate the power of newly released features in torch.

When I want to run llama2 model with 128k long sequence, how can we enable it? I have some experience with DeepSpeed-Ulysses, so the question becomes does torchtitan support sequence parallelism in DeepSpeed-Ulysses?

Thanks!

@gnadathur
Copy link
Contributor

gnadathur commented Nov 15, 2024

@zigzagcai
titan support sequence and context parallelism for ultra long sequence length. cc: @XilunWu , could you share the details on the config.

Oh -- misread this comment, titan supports native FSDP, no deepspeed integration

@XilunWu
Copy link
Contributor

XilunWu commented Nov 15, 2024

Sequence parallel should be enabled by default if TP is used.

To enable CP, please change the context_parallel_degree (see PR #592 description for details/examples)

@tianyu-l tianyu-l added the question Further information is requested label Nov 18, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

4 participants