Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix MSVC build; and associated merge #4

Merged
merged 225 commits into from
Jan 22, 2025
Merged

Conversation

Jeadie
Copy link

@Jeadie Jeadie commented Jan 22, 2025

🗣 Description

🔨 Related Issues

🤔 Concerns

lz1998 and others added 30 commits August 1, 2024 08:26
* bert attention mask

* Allow for using None as a mask.

* Revert part of the changes so that the proper default mask applies.

* Cosmetic change.

* Another cosmetic tweak.

---------

Co-authored-by: Laurent <[email protected]>
* Add updated mfa metallib

* Add bgemm and tests
* fix: fix jina bert example logic

* feat: enable jina embeddings de

* feat: allow more flexibility on Jina Bert
* Fix cargo fmt.

* Clippy fix.

* Cosmetic tweaks.
* Add the flux autoencoder.

* Add the encoder down-blocks.

* Upsampling in the decoder.

* Sketch the flow matching model.

* More flux model.

* Add some of the positional embeddings.

* Add the rope embeddings.

* Add the sampling functions.

* Add the flux example.

* Fix the T5 bits.

* Proper T5 tokenizer.

* Clip encoder path fix.

* Get the clip embeddings.

* No configurable weights in layer norm.

* More weights related fixes.

* Yet another shape fix.

* DType fix.

* Fix a couple more shape issues.

* DType fixes.

* Fix the latent dims.

* Fix more shape issues.

* Autoencoder fixes.

* Get some generations out.

* Bugfix.

* T5 padding.

* Clippy fix.

* Add the decode only mode.

* Fix.

* More fixes.

* Finally get some generations to work.

* Add readme.
* add models support and example for THUDM/glm-4

* fix the ci report

* fmt

* fix

* Update README.org

* Update README.org

* fmt

* Update README.org

* README.md add codegeex4

* README.md add glm4

* Typo.

* change expect into ?

---------

Co-authored-by: Laurent Mazare <[email protected]>
EricLBuehler and others added 26 commits November 12, 2024 20:18
* Add attn softmax

* Add some docs

* Add bf16

* Update kernels for f16

* All tests pass

* Fix cpu clippy

* Fix doc
* Test passes

* All tests pass

* Now all the tests really pass

* Try out always using mm

* Mirror llama.cpp metric

* Mirror llama.cpp metric

* Update test
* Don't always compile fp8

* Correct includes
* Add bf16, f32 conversions for f8e4m3 on metal

* Support storage from slice
* Metal fast sdpa for long seqlen

* Add a test

* Add softcapping test
* General Metal bf16 support

* Fix compilation

* Fix compilation

* Math

* Math unary

* Math affine

* Maybe diambiguate

* Define hugevalbf

* Define matrix

* Pare back?

* Fix

* Format

* Clippy

* Clippy
* Add q4k quantization with imatrix

* Sketch some imatrix generation

* Fixes

* Add quantize_imatrix_onto

* Support loading the imatrix file

* Fix load_imatrix

* Implement imatrix quantization for q2k

* Implement imatrix quantization for q3k

* Fix build on cuda

* Add imatrix q5k, q6k quants
* Add inplace softmax

* inplace_attn_softmax_last_dim

* Fix cuda

* Format
* Metal addmm support

* Format
* Begin to remove ug

* Begin to remove ug

* Begin to remove ug

* Begin to remove ug
@Jeadie Jeadie self-assigned this Jan 22, 2025
@Jeadie Jeadie merged commit 296f14f into spiceai Jan 22, 2025
15 of 22 checks passed
@sgrebnov
Copy link

sgrebnov commented Jan 22, 2025

@Jeadie - I think moving forward instead of merging upstream to spiceai we should create new spiceai-0.8.2 branch based on upstream version / tag and propagate our specific fixes - ideally this should be none (everything should be upstreamed and release as part of official candle library)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.