BharatSahAIyak / autotune

A comprehensive toolkit for seamless data generation and fine-tuning of NLP models, all conveniently packed into a single block.
MIT License
9 stars 5 forks source link

feat: bge/embedding models fine-tuning addition #140

Closed kartikbhtt7 closed 2 months ago

kartikbhtt7 commented 3 months ago

Summary by CodeRabbit

coderabbitai[bot] commented 3 months ago
Walkthrough ## Walkthrough The recent updates enhance the project's functionality and adaptability in handling machine learning tasks, particularly focusing on embedding fine-tuning. Key changes include upgrading dependencies, introducing a new `EmbeddingFineTuning` class for managing fine-tuning tasks, and refining task management logic throughout various modules. These enhancements improve the handling of datasets, model training, and evaluation processes. ## Changes | Files | Change Summary | |---------------------------------------------------|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | `pyproject.toml` | Upgraded `accelerate` from `^0.27.2` to `^0.33.0`, removed `ragatouille`, and added new dependencies (`sentence-transformers`, `catalogue`, etc.) for enhanced functionality and performance. | | `workflow/mixins.py`, `workflow/training/train.py` | Expanded the `dispatch` method to include support for `"embedding_finetuning"` in addition to `"whisper_finetuning"` and modified the `upload_cache` function to exclude both tasks from caching operations. | | `workflow/serializers.py` | Added an optional `guide_model` field to the `ModelDataSerializer` class to capture guide model names for embedding fine-tuning tasks. | | `workflow/training/embedding_finetuning.py` | Introduced a new `EmbeddingFineTuning` class for fine-tuning sentence embedding models, incorporating dataset loading, preparation, training arguments, and model hub uploading functionalities. | | `workflow/training/onnx_conversion.py` | Updated the `convert_to_onnx` function to rename tasks from `"embedding"` to `"embedding_finetuning"`, refining task specificity for fine-tuning operations. | | `workflow/training/utils.py` | Added support for "embedding_finetuning" in the `get_task_class` and `get_model_class` functions, enhancing task recognition and model retrieval capabilities. | | `workflow/utils.py` | Introduced new task configuration and mapping for "embedding_finetuning" in the `get_task_config` and `get_task_mapping` functions, allowing structured management of the associated tasks and data. |

[!TIP]

Announcements - The review status is no longer posted as a separate comment when there are no actionable or nitpick comments. In such cases, the review status is included in the walkthrough comment. - We have updated our review workflow to use the Anthropic's Claude family of models. Please share any feedback in the [discussion post](https://discordapp.com/channels/1134356397673414807/1279579842131787838) on our Discord. - Possibly related PRs: Walkthrough comment now includes a list of potentially related PRs to help you recall past context. Please share any feedback in the [discussion post](https://discordapp.com/channels/1134356397673414807/1282535539299323995) on our Discord. - Suggested labels: CodeRabbit can now suggest labels by learning from your past PRs in the walkthrough comment. You can also provide custom labeling instructions in the UI or configuration file. - Possibly related PRs, automatic label suggestions based on past PRs, learnings, and possibly related issues require data opt-in (enabled by default).

Recent review details **Configuration used: CodeRabbit UI** **Review profile: CHILL**
Commits Files that changed from the base of the PR and between faa3fea47ffe3429ccdfe8d24880cc23a15c0bfb and b902a4357a10cf8244772d36cb282be0f10c661b.
Files selected for processing (1) * workflow/training/embedding_finetuning.py (1 hunks)
Additional comments not posted (6)
workflow/training/embedding_finetuning.py (6)
`12-15`: **LGTM!** The code changes are approved. --- `17-33`: **LGTM!** The code changes are approved. --- `35-38`: **LGTM!** The code changes are approved. --- `40-50`: **LGTM!** The code changes are approved. --- `67-83`: **LGTM!** The code changes are approved. --- `85-88`: **LGTM!** The code changes are approved.
--- Thank you for using CodeRabbit. We offer it for free to the OSS community and would appreciate your support in helping us grow. If you find it useful, would you consider giving us a shout-out on your favorite social media?
Share - [X](https://twitter.com/intent/tweet?text=I%20just%20used%20%40coderabbitai%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20the%20proprietary%20code.%20Check%20it%20out%3A&url=https%3A//coderabbit.ai) - [Mastodon](https://mastodon.social/share?text=I%20just%20used%20%40coderabbitai%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20the%20proprietary%20code.%20Check%20it%20out%3A%20https%3A%2F%2Fcoderabbit.ai) - [Reddit](https://www.reddit.com/submit?title=Great%20tool%20for%20code%20review%20-%20CodeRabbit&text=I%20just%20used%20CodeRabbit%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20proprietary%20code.%20Check%20it%20out%3A%20https%3A//coderabbit.ai) - [LinkedIn](https://www.linkedin.com/sharing/share-offsite/?url=https%3A%2F%2Fcoderabbit.ai&mini=true&title=Great%20tool%20for%20code%20review%20-%20CodeRabbit&summary=I%20just%20used%20CodeRabbit%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20proprietary%20code)
Tips ### Chat There are 3 ways to chat with [CodeRabbit](https://coderabbit.ai): - Review comments: Directly reply to a review comment made by CodeRabbit. Example: - `I pushed a fix in commit .` - `Generate unit testing code for this file.` - `Open a follow-up GitHub issue for this discussion.` - Files and specific lines of code (under the "Files changed" tab): Tag `@coderabbitai` in a new review comment at the desired location with your query. Examples: - `@coderabbitai generate unit testing code for this file.` - `@coderabbitai modularize this function.` - PR comments: Tag `@coderabbitai` in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples: - `@coderabbitai generate interesting stats about this repository and render them as a table.` - `@coderabbitai show all the console.log statements in this repository.` - `@coderabbitai read src/utils.ts and generate unit testing code.` - `@coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.` - `@coderabbitai help me debug CodeRabbit configuration file.` Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments. ### CodeRabbit Commands (Invoked using PR comments) - `@coderabbitai pause` to pause the reviews on a PR. - `@coderabbitai resume` to resume the paused reviews. - `@coderabbitai review` to trigger an incremental review. This is useful when automatic reviews are disabled for the repository. - `@coderabbitai full review` to do a full review from scratch and review all the files again. - `@coderabbitai summary` to regenerate the summary of the PR. - `@coderabbitai resolve` resolve all the CodeRabbit review comments. - `@coderabbitai configuration` to show the current CodeRabbit configuration for the repository. - `@coderabbitai help` to get help. ### Other keywords and placeholders - Add `@coderabbitai ignore` anywhere in the PR description to prevent this PR from being reviewed. - Add `@coderabbitai summary` to generate the high-level summary at a specific location in the PR description. - Add `@coderabbitai` anywhere in the PR title to generate the title automatically. ### CodeRabbit Configuration File (`.coderabbit.yaml`) - You can programmatically configure CodeRabbit by adding a `.coderabbit.yaml` file to the root of your repository. - Please see the [configuration documentation](https://docs.coderabbit.ai/guides/configure-coderabbit) for more information. - If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: `# yaml-language-server: $schema=https://coderabbit.ai/integrations/schema.v2.json` ### Documentation and Community - Visit our [Documentation](https://coderabbit.ai/docs) for detailed information on how to use CodeRabbit. - Join our [Discord Community](https://discord.com/invite/GsXnASn26c) to get help, request features, and share feedback. - Follow us on [X/Twitter](https://twitter.com/coderabbitai) for updates and announcements.