cvat-ai / cvat

Annotate better with CVAT, the industry-leading data engine for machine learning. Used and trusted by teams at any scale, for data of any scale.
https://cvat.ai
MIT License
12.62k stars 3.01k forks source link

Export only annotated images option. #8340

Open bschultz96 opened 2 months ago

bschultz96 commented 2 months ago

Added radio buttons to export either all images or only annotated images when "Save Images" is selected during export process.

The default is set to True to match with the previous behavior when saving images.

Motivation and context

This is a useful addition when creating and adding to datasets. We had been using a work around to export only the annotated images but decided doing the work to implement this would be worthwhile.

How has this been tested?

Manual testing, exporting projects, tasks, and jobs have all been tested.

Checklist

License

Summary by CodeRabbit

coderabbitai[bot] commented 2 months ago

[!IMPORTANT]

Review skipped

Auto incremental reviews are disabled on this repository.

Please check the settings in the CodeRabbit UI or the .coderabbit.yaml file in this repository. To trigger a single review, invoke the @coderabbitai review command.

You can disable this status message by setting the reviews.review_status to false in the CodeRabbit configuration file.

Walkthrough

The recent changes across the codebase introduce a new parameter, allImages, to various export-related functions and classes. This parameter allows users to specify whether all images should be included in dataset exports. Modifications were made to function signatures, internal logic, and API documentation to accommodate this new functionality, improving the granularity and flexibility of the export processes throughout the application.

Changes

Files Change Summary
cvat-core/src/annotations.ts, cvat-core/src/project-implementation.ts, cvat-core/src/project.ts, cvat-core/src/server-proxy.ts, cvat-core/src/session-implementation.ts, cvat-core/src/session.ts, cvat-ui/src/actions/export-actions.ts Added allImages parameter to export functions, modifying function signatures and internal logic to include this parameter for more control over image inclusion during exports.
cvat-ui/src/components/export-dataset/export-dataset-modal.tsx Updated form in the export dataset modal to include a new option for saving all images, modifying the FormValues type and handling state management accordingly.
cvat/apps/dataset_manager/bindings.py, cvat/apps/dataset_manager/formats/*.py Introduced all_images parameter in multiple functions and classes, enhancing the ability to control image inclusion in dataset exports across various formats.
cvat/apps/dataset_manager/project.py, cvat/apps/dataset_manager/task.py Updated export_project, export_job, and export_task functions to include the all_images parameter in their signatures, allowing for more dynamic export options.
cvat/apps/dataset_manager/util.py, cvat/apps/engine/background.py, cvat/apps/engine/views.py Modified functions to integrate the all_images parameter for enhanced export filename generation and callback handling, improving the granularity of image export control across the engine.
cvat/schema.yml Added a new query parameter all_images as a boolean to allow users to specify whether to include all images in dataset exports.

Sequence Diagram(s)

sequenceDiagram
    participant User
    participant API
    participant ExportManager
    participant Dataset

    User->>API: Request Export with all_images
    API->>ExportManager: Pass all_images parameter
    ExportManager->>Dataset: Export Dataset with all_images
    Dataset-->>ExportManager: Return Export Status
    ExportManager-->>API: Return API Response
    API-->>User: Provide Export Results

🐇 In the meadow, I hop and play,
With new exports brightening the day!
All images saved, what a delight,
A dataset journey, pure and bright!
Let's celebrate this joyous feat,
With carrots and love, life is sweet! 🥕✨


Thank you for using CodeRabbit. We offer it for free to the OSS community and would appreciate your support in helping us grow. If you find it useful, would you consider giving us a shout-out on your favorite social media?

Share - [X](https://twitter.com/intent/tweet?text=I%20just%20used%20%40coderabbitai%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20the%20proprietary%20code.%20Check%20it%20out%3A&url=https%3A//coderabbit.ai) - [Mastodon](https://mastodon.social/share?text=I%20just%20used%20%40coderabbitai%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20the%20proprietary%20code.%20Check%20it%20out%3A%20https%3A%2F%2Fcoderabbit.ai) - [Reddit](https://www.reddit.com/submit?title=Great%20tool%20for%20code%20review%20-%20CodeRabbit&text=I%20just%20used%20CodeRabbit%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20proprietary%20code.%20Check%20it%20out%3A%20https%3A//coderabbit.ai) - [LinkedIn](https://www.linkedin.com/sharing/share-offsite/?url=https%3A%2F%2Fcoderabbit.ai&mini=true&title=Great%20tool%20for%20code%20review%20-%20CodeRabbit&summary=I%20just%20used%20CodeRabbit%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20proprietary%20code)
Tips ### Chat There are 3 ways to chat with [CodeRabbit](https://coderabbit.ai): - Review comments: Directly reply to a review comment made by CodeRabbit. Example: - `I pushed a fix in commit .` - `Generate unit testing code for this file.` - `Open a follow-up GitHub issue for this discussion.` - Files and specific lines of code (under the "Files changed" tab): Tag `@coderabbitai` in a new review comment at the desired location with your query. Examples: - `@coderabbitai generate unit testing code for this file.` - `@coderabbitai modularize this function.` - PR comments: Tag `@coderabbitai` in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples: - `@coderabbitai generate interesting stats about this repository and render them as a table.` - `@coderabbitai show all the console.log statements in this repository.` - `@coderabbitai read src/utils.ts and generate unit testing code.` - `@coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.` - `@coderabbitai help me debug CodeRabbit configuration file.` Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments. ### CodeRabbit Commands (Invoked using PR comments) - `@coderabbitai pause` to pause the reviews on a PR. - `@coderabbitai resume` to resume the paused reviews. - `@coderabbitai review` to trigger an incremental review. This is useful when automatic reviews are disabled for the repository. - `@coderabbitai full review` to do a full review from scratch and review all the files again. - `@coderabbitai summary` to regenerate the summary of the PR. - `@coderabbitai resolve` resolve all the CodeRabbit review comments. - `@coderabbitai configuration` to show the current CodeRabbit configuration for the repository. - `@coderabbitai help` to get help. ### Other keywords and placeholders - Add `@coderabbitai ignore` anywhere in the PR description to prevent this PR from being reviewed. - Add `@coderabbitai summary` to generate the high-level summary at a specific location in the PR description. - Add `@coderabbitai` anywhere in the PR title to generate the title automatically. ### CodeRabbit Configuration File (`.coderabbit.yaml`) - You can programmatically configure CodeRabbit by adding a `.coderabbit.yaml` file to the root of your repository. - Please see the [configuration documentation](https://docs.coderabbit.ai/guides/configure-coderabbit) for more information. - If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: `# yaml-language-server: $schema=https://coderabbit.ai/integrations/schema.v2.json` ### Documentation and Community - Visit our [Documentation](https://coderabbit.ai/docs) for detailed information on how to use CodeRabbit. - Join our [Discord Community](https://discord.com/invite/GsXnASn26c) to get help, request features, and share feedback. - Follow us on [X/Twitter](https://twitter.com/coderabbitai) for updates and announcements.
zhiltsov-max commented 2 months ago

This is a useful addition when creating and adding to datasets.

If you have some automation for task creation, so you know was it a dataset originally, you can upload only empty images or upload everything and mark the extra images as deleted.

bschultz96 commented 2 months ago

Please add at least some tests for the updated functionality.

I'm getting a bunch of tests failing due to an APIException with the following error message: "TypeError: _export_task_orjob() got an unexpected keyword argument all_images." It traces back to the cvat.py format file which does indeed have the parameter all_images within all necessary functions. Interestingly enough, if I change the test to use a different format it works fine. Just wondering if you might know why specifically for cvat format I get this exception? Screenshot linked with the output I get.

Screenshot 2024-08-28 100947

zhiltsov-max commented 2 months ago

Maybe you need to rebuild the containers before running tests.

KTXKIKI commented 2 months ago

@ bschultz96 Hello brother, have you made any progress? Perhaps there are other ideas for implementation?

bschultz96 commented 2 months ago

@ bschultz96 Hello brother, have you made any progress? Perhaps there are other ideas for implementation?

I couldn't figure out why the tests I added to test the new functionality aren't returning the results I expect. When I run the tests the endpoint is always downloading either all or no images and behavior doesn't change for whichever all_images is set to. Manually testing the endpoint works fine, so I think I'm missing something in the testing environment backend but after spending a few hours couldn't find make any headway.

zhiltsov-max commented 2 months ago

I couldn't figure out why the tests I added to test the new functionality aren't returning the results I expect. When I run the tests the endpoint is always downloading either all or no images and behavior doesn't change for whichever all_images is set to. Manually testing the endpoint works fine, so I think I'm missing something in the testing environment backend but after spending a few hours couldn't find make any headway.

Have you tried pytest --rebuild tests/python?

sonarcloud[bot] commented 2 months ago

Quality Gate Passed Quality Gate passed

Issues
0 New issues
0 Accepted issues

Measures
0 Security Hotspots
0.0% Coverage on New Code
0.0% Duplication on New Code

See analysis details on SonarCloud