-
Notifications
You must be signed in to change notification settings - Fork 167
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add new inference paradiam #473
base: main
Are you sure you want to change the base?
Conversation
273fcd9
to
03e0ca7
Compare
03e0ca7
to
f48d1b1
Compare
e0641e2
to
5c6fd44
Compare
2a45262
to
48b6adc
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
In this Era, it is critical and urgent to introduce LLM in Sedna. The new paradigm shown in this proposal is thus important.
As discussed, since the Sedna lib is removed from the paradigm, we need several serious discussions on this proposal.
-
Sedna and Ianvs paradigms for LLM joint inference should be consistent - might discussed with @FuryMartin and @hsj576
-
How will this new paradigm improve the integration of LLM models by removing Sedna lib in cloud and edge inference?
- Will that bring concerns for the data interfaces? When switching to a new model, people need to switch the data format to the new, leading to an additional cost.
- Will this paradigm support a non-LLM model on the edge? It seems that this paradigm only works in the case where both cloud and edge deploy LLMs, instead of cloud-only LLM and edge-only LLM, as claimed in the proposal.
- How will this new paradigm improve the integration of hard example mining by removing Sedna lib in hard example mining (sider car)?
- What is the hard example mining algorithm used now in the new paradigm? How to switch hard example mining algorithms?
- With the above consideration, we wonder whether this paradigm fully supports LLM joint inference. The new paradigm seems more like a new paradigm of LLM single-task learning, doesn't it? That is, a trained LLM is deployed either on the edge or on the cloud.
Signed-off-by: ming.tang <[email protected]>
48b6adc
to
1fc1629
Compare
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
What type of PR is this?
/kind documentation
What this PR does / why we need it:
The current joint inference paradigm can only support the traditional discriminative model and cannot support the large language model (LLM). This proposal aims to introduce a new inference paradigm to support the joint inference of LLM
Which issue(s) this PR fixes:
Fixes #