Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llama3.1 #56

Merged
merged 18 commits into from
Jul 25, 2024
Merged

Llama3.1 #56

merged 18 commits into from
Jul 25, 2024

Conversation

farzadab
Copy link
Contributor

@farzadab farzadab commented Jul 25, 2024

This PR enables Llama 3.1 training and loading.

Using Llama 3.1 required us to upgrade transformers. The new version changes caching so there are some changes related to that as well.

@farzadab farzadab changed the title [WIP] Llama3.1 Llama3.1 Jul 25, 2024
@farzadab farzadab marked this pull request as ready for review July 25, 2024 18:28
@farzadab farzadab requested review from juberti and zqhuang211 July 25, 2024 18:29
@farzadab
Copy link
Contributor Author

Huh, the transformers version change was note pushed.

@zqhuang211 zqhuang211 self-requested a review July 25, 2024 21:49
@farzadab farzadab merged commit 01960b1 into main Jul 25, 2024
1 check passed
@farzadab farzadab deleted the farzad-llama3.1 branch July 25, 2024 23:57
akshat0311 pushed a commit to jiviai/audio-llm that referenced this pull request Jan 30, 2025
* upgrade transformers for llama 3.1
* fix cache on new version
* TRAIN_ARGS instead of UV_CONFIG
* remove matchtrain validation
zqhuang211 added a commit that referenced this pull request Feb 12, 2025
- add fleurs data set
- update WER calculation to use whisper normalization to be consistent with the whisper paper
- update several datasets to include proper lang_id that is needed for whisper-normalization
- add new eval configs for commonvoice, fleurs, and covost2
- add test cases for wer and bleu
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants