Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add support for anthropic, azure, aleph alpha, ai21, togetherai, cohere, replicate, huggingface inference endpoints, etc. #660

Merged
merged 3 commits into from
Sep 7, 2023

Conversation

krrishdholakia
Copy link
Contributor

Hi @AntonOsika,

Following up on this PR - #574.

I fixed the streaming bugs and confirmed it works with azure.

Here is the gpt-engineer working with my deployed azure instance:
Screenshot 2023-09-02 at 4 26 21 PM

Here is the gpt-engineer working with ai21's "j2-mid" model:
Screenshot 2023-09-02 at 4 26 35 PM

For those trying to use WizardCoder / Phind-CodeLlama, we (+ this PR) also provides support for Huggingface Inference Endpoints and Baseten.
https://docs.litellm.ai/docs/completion/supported

If anyone's using gpt-engineer in production, and would like to split traffic between a finetuned model and gpt-4 - they can do that too: https://docs.litellm.ai/docs/tutorials/ab_test_llms

Please let me know if this PR looks good or you'd still prefer to stash it.

Happy to make any changes / update docs, if the initial PR looks good.

@AntonOsika
Copy link
Owner

@ErikBjare what do you think of this?

It feels like it's too much have both:

  • API_BASE_URL
  • Azure flags
  • LiteLLM

We should simplify and pick a subset (maybe just one?) of all of these!

@krrishdholakia
Copy link
Contributor Author

@AntonOsika happy to make any changes as necessary 🙂

@AntonOsika
Copy link
Owner

AntonOsika commented Sep 7, 2023

Thought about it and I think we should go for it! @krrishdholakia

Do we still need the special azure parameters after this? Would be great to have that taken care of by env variables + litellm instead of code bloat! 🚀

@AntonOsika AntonOsika merged commit 9079f84 into AntonOsika:main Sep 7, 2023
AntonOsika added a commit that referenced this pull request Sep 7, 2023
…ai, cohere, replicate, huggingface inference endpoints, etc. (#660)"

This reverts commit 9079f84.
AntonOsika added a commit that referenced this pull request Sep 7, 2023
…ai, cohere, replicate, huggingface inference endpoints, etc. (#660)" (#685)

This reverts commit 9079f84.
@AntonOsika
Copy link
Owner

Hey @krrishdholakia I got this and reverted:
image

Do you have any ideas why?

@krrishdholakia
Copy link
Contributor Author

What version of langchain were you using?

@krrishdholakia
Copy link
Contributor Author

Hey @AntonOsika here's chatlitellm working for replicate and cohere
Screenshot 2023-09-07 at 9 23 49 PM

Here it is for OpenAI + Anthropic
Screenshot 2023-09-07 at 9 24 15 PM

Can you please let me know the version you're using so i can debug this?

70ziko pushed a commit to 70ziko/gpt-engineer that referenced this pull request Oct 25, 2023
…re, replicate, huggingface inference endpoints, etc. (AntonOsika#660)

* fix streaming bug

* fixing streaming bug and adding litellm to pyproject.toml

* adding provider details to env template
70ziko pushed a commit to 70ziko/gpt-engineer that referenced this pull request Oct 25, 2023
…ai, cohere, replicate, huggingface inference endpoints, etc. (AntonOsika#660)" (AntonOsika#685)

This reverts commit 9079f84.
@anushrxy
Copy link

anushrxy commented Dec 2, 2023

Hi @AntonOsika,

Following up on this PR - #574.

I fixed the streaming bugs and confirmed it works with azure.

Here is the gpt-engineer working with my deployed azure instance: Screenshot 2023-09-02 at 4 26 21 PM

Here is the gpt-engineer working with ai21's "j2-mid" model: Screenshot 2023-09-02 at 4 26 35 PM

For those trying to use WizardCoder / Phind-CodeLlama, we (+ this PR) also provides support for Huggingface Inference Endpoints and Baseten. https://docs.litellm.ai/docs/completion/supported

If anyone's using gpt-engineer in production, and would like to split traffic between a finetuned model and gpt-4 - they can do that too: https://docs.litellm.ai/docs/tutorials/ab_test_llms

Please let me know if this PR looks good or you'd still prefer to stash it.

Happy to make any changes / update docs, if the initial PR looks good.

How do someone use GPT 3.5 turbo Fine tuned models in it?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants