Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix FPEBC train pipeline test #2090

Closed
wants to merge 2 commits into from
Closed

Conversation

sarckk
Copy link
Member

@sarckk sarckk commented Jun 8, 2024

Summary:
3 issues that needed fixing:

  1. Move batch to GPU
  2. Set compute kernel to fused instead of dense to work w/ TW sharding
  3. Ensure that input batch idlist_features KJT has max length equal to the max lengths specified for feature processors (otherwise it would fail on torch.gather() in feature processor due to shape mismatch between KJT input lengths and indices

Differential Revision: D56950454

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Jun 8, 2024
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D56950454

@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D56950454

sarckk added a commit to sarckk/torchrec that referenced this pull request Jun 10, 2024
Summary:
Pull Request resolved: pytorch#2090

3 issues that needed fixing:
1) Move batch to GPU
2) Set compute kernel to fused instead of dense to work w/ TW sharding
3) Ensure that input batch idlist_features KJT has max length equal to the max lengths specified for feature processors (otherwise it would fail on `torch.gather()` in feature processor due to  shape mismatch between KJT input lengths and indices

Reviewed By: henrylhtsang

Differential Revision: D56950454
@sarckk sarckk force-pushed the export-D56950454 branch from caf3723 to 0fe2fef Compare June 10, 2024 15:28
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D56950454

sarckk added a commit to sarckk/torchrec that referenced this pull request Jun 10, 2024
Summary:
Pull Request resolved: pytorch#2090

3 issues that needed fixing:
1) Move batch to GPU
2) Set compute kernel to fused instead of dense to work w/ TW sharding
3) Ensure that input batch idlist_features KJT has max length equal to the max lengths specified for feature processors (otherwise it would fail on `torch.gather()` in feature processor due to  shape mismatch between KJT input lengths and indices

Reviewed By: henrylhtsang

Differential Revision: D56950454
@sarckk sarckk force-pushed the export-D56950454 branch from 0fe2fef to bb15946 Compare June 10, 2024 17:34
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D56950454

sarckk added a commit to sarckk/torchrec that referenced this pull request Jun 10, 2024
Summary:
Pull Request resolved: pytorch#2090

3 issues that needed fixing:
1) Move batch to GPU
2) Set compute kernel to fused instead of dense to work w/ TW sharding
3) Ensure that input batch idlist_features KJT has max length equal to the max lengths specified for feature processors (otherwise it would fail on `torch.gather()` in feature processor due to  shape mismatch between KJT input lengths and indices

Reviewed By: henrylhtsang

Differential Revision: D56950454
@sarckk sarckk force-pushed the export-D56950454 branch from bb15946 to a80d5c2 Compare June 10, 2024 19:40
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D56950454

@sarckk sarckk force-pushed the export-D56950454 branch from a80d5c2 to 8489857 Compare June 11, 2024 02:04
sarckk added a commit to sarckk/torchrec that referenced this pull request Jun 11, 2024
Summary:
Pull Request resolved: pytorch#2090

3 issues that needed fixing:
1) Move batch to GPU
2) Set compute kernel to fused instead of dense to work w/ TW sharding
3) Ensure that input batch idlist_features KJT has max length equal to the max lengths specified for feature processors (otherwise it would fail on `torch.gather()` in feature processor due to  shape mismatch between KJT input lengths and indices

Reviewed By: henrylhtsang

Differential Revision: D56950454
Summary:
Pull Request resolved: pytorch#2090

3 issues that needed fixing:
1) Move batch to GPU
2) Set compute kernel to fused instead of dense to work w/ TW sharding
3) Ensure that input batch idlist_features KJT has max length equal to the max lengths specified for feature processors (otherwise it would fail on `torch.gather()` in feature processor due to  shape mismatch between KJT input lengths and indices

Reviewed By: henrylhtsang

Differential Revision: D56950454
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D56950454

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants