Bumps the pip group with 2 updates in the /Experiments/CV/ocr_with_bert directory: transformers and pydantic.
Bumps the pip group with 1 update in the /MachineLearning/GMM/GMM_phones directory: scikit-learn.
Llama 3 is supported in this release through the Llama 2 architecture and some fixes in the tokenizers library.
Idefics2
The Idefics2 model was created by the Hugging Face M4 team and authored by Léo Tronchon, Hugo Laurencon, Victor Sanh. The accompanying blog post can be found here.
Idefics2 is an open multimodal model that accepts arbitrary sequences of image and text inputs and produces text outputs. The model can answer questions about images, describe visual content, create stories grounded on multiple images, or simply behave as a pure language model without visual inputs. It improves upon IDEFICS-1, notably on document understanding, OCR, or visual reasoning. Idefics2 is lightweight (8 billion parameters) and treats images in their native aspect ratio and resolution, which allows for varying inference efficiency.
Recurrent Gemma architecture. Taken from the original paper.
The Recurrent Gemma model was proposed in RecurrentGemma: Moving Past Transformers for Efficient Open Language Models by the Griffin, RLHF and Gemma Teams of Google.
The abstract from the paper is the following:
We introduce RecurrentGemma, an open language model which uses Google’s novel Griffin architecture. Griffin combines linear recurrences with local attention to achieve excellent performance on language. It has a fixed-sized state, which reduces memory use and enables efficient inference on long sequences. We provide a pre-trained model with 2B non-embedding parameters, and an instruction tuned variant. Both models achieve comparable performance to Gemma-2B despite being trained on fewer tokens.
Jamba is a pretrained, mixture-of-experts (MoE) generative text model, with 12B active parameters and an overall of 52B parameters across all experts. It supports a 256K context length, and can fit up to 140K tokens on a single 80GB GPU.
As depicted in the diagram below, Jamba’s architecture features a blocks-and-layers approach that allows Jamba to successfully integrate Transformer and Mamba architectures altogether. Each Jamba block contains either an attention or a Mamba layer, followed by a multi-layer perceptron (MLP), producing an overall ratio of one Transformer layer out of every eight total layers.
Jamba introduces the first HybridCache object that allows it to natively support assisted generation, contrastive search, speculative decoding, beam search and all of the awesome features from the generate API!
Fixes the maxlen property being dropped on deque validation. Happened only if the deque item has been typed. Changes the _validate_sequence_like func, #6581 by @maciekglowka
We're happy to announce the 0.23.1 release which fixes a few issues affecting many users, namely: K-Means should be faster for small sample sizes, and the representation of third-party estimators was fixed.
We're happy to announce the 0.22.2.post1 bugfix release.
The 0.22.2.post1 release includes a packaging fix for the source distribution
but the content of the packages is otherwise identical to the content of the
wheels with the 0.22.2 version (without the .post1 suffix).
Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR:
- `@dependabot rebase` will rebase this PR
- `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it
- `@dependabot merge` will merge this PR after your CI passes on it
- `@dependabot squash and merge` will squash and merge this PR after your CI passes on it
- `@dependabot cancel merge` will cancel a previously requested merge and block automerging
- `@dependabot reopen` will reopen this PR if it is closed
- `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually
- `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency
- `@dependabot ignore major version` will close this group update PR and stop Dependabot creating any more for the specific dependency's major version (unless you unignore this specific dependency's major version or upgrade to it yourself)
- `@dependabot ignore minor version` will close this group update PR and stop Dependabot creating any more for the specific dependency's minor version (unless you unignore this specific dependency's minor version or upgrade to it yourself)
- `@dependabot ignore ` will close this group update PR and stop Dependabot creating any more for the specific dependency (unless you unignore this specific dependency or upgrade to it yourself)
- `@dependabot unignore ` will remove all of the ignore conditions of the specified dependency
- `@dependabot unignore ` will remove the ignore condition of the specified dependency and ignore conditions
You can disable automated security fix PRs for this repo from the [Security Alerts page](https://github.com/YeonwooSung/ai_book/network/alerts).
Bumps the pip group with 2 updates in the /Experiments/CV/ocr_with_bert directory: transformers and pydantic. Bumps the pip group with 1 update in the /MachineLearning/GMM/GMM_phones directory: scikit-learn.
Updates
transformers
from 4.39.3 to 4.40.1Release notes
Sourced from transformers's releases.
... (truncated)
Commits
9fe3f58
v4.40.1f8fec6b
Make EosTokenCriteria compatible with mps (#30376)745bbfe
Release: v4.40.05728b5a
FIX: Fixes unexpected behaviour for Llava / LLama & AWQ Fused modules + rever...005b957
Add DBRX Model (#29921)63c5e27
Do not drop mask with SDPA for more cases (#30311)acab997
Revert "Re-enable SDPA's FA2 path (#30070)" (#30314)7509a0a
Fix RecurrentGemma device_map (#30273)9459efb
Add atol for sliding window test (#30303)3f20877
Add jamba (#29943)Updates
pydantic
from 1.10.11 to 1.10.13Release notes
Sourced from pydantic's releases.
Changelog
Sourced from pydantic's changelog.
Commits
8822578
Prepare release 1.10.13 (#7674)59d8f38
[Backport] Add max length check tovalidate_email
(#7673)69b92b5
Make the v1 mypy plugin work with both v1 and v2 (#6920)87bf417
Update pip commands to install 1.10 (#6930)d9c2af3
Prepare release 1.10.12 (#6825)2aaddf6
Deque's maxlen property dropped on V1 validation (#6586)Updates
scikit-learn
from 0.19.1 to 0.23.1Release notes
Sourced from scikit-learn's releases.
... (truncated)
Commits
fd23727
MNT Completes position arg deprecation (#17272)467d95f
DOC Add threadpoolctl requirement to documentation (#17273)256afb3
DOC prepare whats_new for 0.23.1 (#17270)3905fc0
MNT bump version to 0.23.12431ea3
BUG Fixes cython code for ppc arch (#17201)04e485e
DOC avoid FutureWarnings for deprecations examples (#17264)1dfc42a
DOC Fix parametrize_with_checks link (#17263)25e6fd1
DOC Add explanation of why iterative imputer is experimental (#17115)6088842
DOC Update plot_column_transformer to notebook style (#17028)c19aa54
MNT Fix incorrect source code link for wrapped objects (#17247)Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting
@dependabot rebase
.Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show