fix(deps): update dependency bitsandbytes to ^0.44.0 #169
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR contains the following updates:
^0.43.1
->^0.44.0
Release Notes
TimDettmers/bitsandbytes (bitsandbytes)
v0.44.1
Compare Source
What's Changed
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.44.0...0.44.1
v0.44.0
: : New AdEMAMix optimizer, Embeddings quantization, and more!Compare Source
New optimizer: AdEMAMix
The AdEMAMix optimizer is a modification to AdamW which proposes tracking two EMAs to better leverage past gradients. This allows for faster convergence with less training data and improved resistance to forgetting.
We've implemented 8bit and paged variations:
AdEMAMix
,AdEMAMix8bit
,PagedAdEMAMix
, andPagedAdEMAMix8bit
. These can be used with a similar API to existing optimizers.8-bit Optimizers Update
The block size for all 8-bit optimizers has been reduced from 2048 to 256 in this release. This is a change from the original implementation proposed in the paper which improves accuracy.
CUDA Graphs support
A fix to enable CUDA Graphs capture of kernel functions was made in #1330. This allows for performance improvements with inference frameworks like vLLM. Thanks @jeejeelee!
Quantization for Embeddings
The trend of LLMs to use larger vocabularies continues. The embeddings can take up a significant portion of a quantized model's footprint. We now have an implementation of
Embedding4bit
andEmbedding8bit
thanks to @galqiwi!Example usage:
Continuous Builds
We are now building binary wheels for each change on
main
. These builds can be used to preview upcoming changes.🚤 Continuous Build
What's Changed
move_to_device
kwarg to the optimizer'sload_state_dict
by @koute in https://github.com/bitsandbytes-foundation/bitsandbytes/pull/1344New Contributors
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.43.3...v0.44.0
Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
This PR was generated by Mend Renovate. View the repository job log.