Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

💡 idea: Soften the Enterprise requirement for multiple LLM backends (Add AI Service) #169

Open
AnnoyingTechnology opened this issue Apr 23, 2024 · 4 comments
Labels
enhancement New feature or request

Comments

@AnnoyingTechnology
Copy link

Description

Screenshot from 2024-04-23 09-42-58

Can the requirement for an Enterprise licence to use multiple LLMs endpoints be challenged ?

If so, maybe allowing 2 to 3 endpoints on community version, and then require Enterprise version would be a nice gesture.

@AnnoyingTechnology AnnoyingTechnology added the enhancement New feature or request label Apr 23, 2024
@esethna
Copy link

esethna commented Apr 29, 2024

Hey @AnnoyingTechnology, thanks for the feedback. Can you help us understand your use case for multi LLM support? Also curious in general how your team is using Mattermost and how large your instance is in terms of user count?

We've heard the overwhelming majority of small teams are using just one LLM model, while larger instances and enterprises may need multi-LLM support to run various custom and tuned models. Also, please check out the design preview of some work we're doing to enhance multi-LLM support even further: #69 (comment). We'd look forward to hearing your feedback on that thread regarding the designs and functionality.

@AnnoyingTechnology
Copy link
Author

AnnoyingTechnology commented Apr 29, 2024

Hi,

Just to answer on the use case :

  • We have roughly ten people who would use this feature, but only 3 on a regular basis.
  • We would use gpt-4-turbo and Mixtral 8x7B on a local 3090.

Most requests

  • we prefer not having the data leave the network
  • we enjoy the low latency ultra high token/s of a local LLM
  • no need for another language than English

Some requests

  • we actually need GPT4
  • and/or we need French language support
  • and/or we need document upload support (currently using ChatGPT for that, but I guess it will come to mattermost)
  • and we are ready to accept that our is data sent to OpenAI

Hence the two LLM endpoints.

That multi-bot feature would be ideal.

@AnnoyingTechnology
Copy link
Author

@esethna I'm bumping this issue from April.

Any chance of reconsideration ?

Currently we would very much like being able to toggle between 4o and Qwen 2.5 32B on-prem.

@esethna
Copy link

esethna commented Oct 15, 2024

Thanks for the feedback. cc// @BillAnderson304 on the above.

I don't believe we have any plans to remove the Enterprise requirement at this time for Copilot Multi-LLM setups, however one things we've heard in the past is people forking the repo and compiling two separate plugins with different bot names and configurations.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants