Holidays season release! 🎄
This Optimum TPU release comes in with a larger support for models, in particular newer Llamas 🦙 for serving and fine-tuning, as well as initial support for the all recent TPU v6e
and few fixes here and there.
What's Changed
- fix(ci): correct TGI_VERSION definition in workflow by @tengomucho in #122
- Fix nightlies again by @tengomucho in #123
- ⚙️ Fix Integration Test for TGI by @baptistecolle in #124
- 🔂 Fix repetition penalty by @tengomucho in #125
- Allow sharding fine tuned misaligned models by @tengomucho in #126
- 🦙 Newer Llamas support by @tengomucho in #129
- 🦙 Add llama fine-tuning notebook example by @baptistecolle in #130
- doc(v6e): mention initial v6e support by @tengomucho in #131
- ⚙️ Refactor TGI Dockerfile to support Google-Cloud-Containers as a target by @baptistecolle in #127
- 🐛 Fix the convergence of loss function for the llama fine tuning example by @baptistecolle in #132
- chore: update version to v0.2.3 by @tengomucho in #133
Full Changelog: v0.2.1...v0.2.3