BharatSahAIyak / autotune

A comprehensive toolkit for seamless data generation and fine-tuning of NLP models, all conveniently packed into a single block.
MIT License
9 stars 5 forks source link

feat: POST /model/iterate #109

Closed sooraj1002 closed 4 months ago

sooraj1002 commented 4 months ago

Summary by CodeRabbit

coderabbitai[bot] commented 4 months ago

[!WARNING]

Review failed

The pull request is closed.

Walkthrough

The recent changes introduce new views and functionalities for handling model iterations and health checks within the autotune module, revamp the data fetcher for language model generation in the workflow generator, and modify several views and utility functions for improved prompt and task configuration. These updates enhance the overall system's capability to manage user requests, generate model examples, and ensure system health efficiently.

Changes

File Path Change Summary
autotune/urls.py Added ModelIterationView, PingCheckView, and HealthCheckView to the URL configurations.
workflow/generator/dataFetcher.py Removed old and added new request_and_save methods with extended parameters. Updated call_llm_generate to include extra parameters for model generation.
workflow/generator/generator_model.py Introduced ModelDataFetcher class with methods for generating and refining examples using a language model, including generate_or_refine, construct_user_prompt, request_and_save, and parse_examples.
workflow/utils.py Modified get_task_config function by updating the user_prompt_template to a new template for sentence classification generation.
workflow/views.py Changed mixins order in TrainModelView and ForceAlignmentView. Added a new class ModelIterationView for custom model iteration logic.

Sequence Diagram(s)

sequenceDiagram
    participant User
    participant URLConf
    participant Views
    participant ModelDataFetcher
    participant DataFetcher

    User->>URLConf: Sends request to Model Iteration URL
    URLConf->>Views: Forwards request to ModelIterationView
    Views->>ModelDataFetcher: Initializes ModelDataFetcher
    ModelDataFetcher->>DataFetcher: Calls request_and_save method
    DataFetcher->>ModelDataFetcher: Returns data
    ModelDataFetcher->>Views: Processes data and constructs response
    Views->>User: Sends response back
sequenceDiagram
    participant User
    participant URLConf
    participant Views

    User->>URLConf: Sends health check request to PingCheckView/HealthCheckView
    URLConf->>Views: Forwards request to appropriate view
    Views->>Views: Performs health check
    Views->>User: Sends health check status

Poem

🐰

In code and scripts, our changes flow, Model iterations, health checks that glow. Fetch the data, parse with might, Train and align, bring models to light. Prompt templates we now redefine, AI grows stronger, line by line.

🚀


[!TIP]

Early access features: enabled We are currently testing the following features in early access: - **OpenAI `gpt-4o` model for code reviews and chat**: OpenAI claims that this model is better at understanding and generating code than the previous models. We seek your feedback over the next few weeks before making it generally available. Note: - You can enable or disable early access features from the CodeRabbit UI or by updating the CodeRabbit configuration file. - Please join our [Discord Community](https://discord.com/invite/GsXnASn26c) to provide feedback and report issues. - OSS projects are currently opted into early access features by default.

Thank you for using CodeRabbit. We offer it for free to the OSS community and would appreciate your support in helping us grow. If you find it useful, would you consider giving us a shout-out on your favorite social media?

Share - [X](https://twitter.com/intent/tweet?text=I%20just%20used%20%40coderabbitai%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20the%20proprietary%20code.%20Check%20it%20out%3A&url=https%3A//coderabbit.ai) - [Mastodon](https://mastodon.social/share?text=I%20just%20used%20%40coderabbitai%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20the%20proprietary%20code.%20Check%20it%20out%3A%20https%3A%2F%2Fcoderabbit.ai) - [Reddit](https://www.reddit.com/submit?title=Great%20tool%20for%20code%20review%20-%20CodeRabbit&text=I%20just%20used%20CodeRabbit%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20proprietary%20code.%20Check%20it%20out%3A%20https%3A//coderabbit.ai) - [LinkedIn](https://www.linkedin.com/sharing/share-offsite/?url=https%3A%2F%2Fcoderabbit.ai&mini=true&title=Great%20tool%20for%20code%20review%20-%20CodeRabbit&summary=I%20just%20used%20CodeRabbit%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20proprietary%20code)
Tips ### Chat There are 3 ways to chat with [CodeRabbit](https://coderabbit.ai): - Review comments: Directly reply to a review comment made by CodeRabbit. Example: - `I pushed a fix in commit .` - `Generate unit testing code for this file.` - `Open a follow-up GitHub issue for this discussion.` - Files and specific lines of code (under the "Files changed" tab): Tag `@coderabbitai` in a new review comment at the desired location with your query. Examples: - `@coderabbitai generate unit testing code for this file.` - `@coderabbitai modularize this function.` - PR comments: Tag `@coderabbitai` in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples: - `@coderabbitai generate interesting stats about this repository and render them as a table.` - `@coderabbitai show all the console.log statements in this repository.` - `@coderabbitai read src/utils.ts and generate unit testing code.` - `@coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.` - `@coderabbitai help me debug CodeRabbit configuration file.` Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments. ### CodeRabbit Commands (invoked as PR comments) - `@coderabbitai pause` to pause the reviews on a PR. - `@coderabbitai resume` to resume the paused reviews. - `@coderabbitai review` to trigger an incremental review. This is useful when automatic reviews are disabled for the repository. - `@coderabbitai full review` to do a full review from scratch and review all the files again. - `@coderabbitai summary` to regenerate the summary of the PR. - `@coderabbitai resolve` resolve all the CodeRabbit review comments. - `@coderabbitai configuration` to show the current CodeRabbit configuration for the repository. - `@coderabbitai help` to get help. Additionally, you can add `@coderabbitai ignore` anywhere in the PR description to prevent this PR from being reviewed. ### CodeRabbit Configration File (`.coderabbit.yaml`) - You can programmatically configure CodeRabbit by adding a `.coderabbit.yaml` file to the root of your repository. - Please see the [configuration documentation](https://docs.coderabbit.ai/guides/configure-coderabbit) for more information. - If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: `# yaml-language-server: $schema=https://coderabbit.ai/integrations/schema.v2.json` ### Documentation and Community - Visit our [Documentation](https://coderabbit.ai/docs) for detailed information on how to use CodeRabbit. - Join our [Discord Community](https://discord.com/invite/GsXnASn26c) to get help, request features, and share feedback. - Follow us on [X/Twitter](https://twitter.com/coderabbitai) for updates and announcements.