Cruiz102 / Advesarial_Attacks_Tests

Adversarial Attacks with Hugginface models.
1 stars 0 forks source link

Add optimizations to the training loop. #1

Open Cruiz102 opened 7 months ago

Cruiz102 commented 7 months ago

To enable efficient training on GPUs and scale our repository for models with millions to billions of parameters—essential for working with large visual language models—we must implement optimization techniques. These include qLora, precision changes, and Flash Attention for Transformers.

To facilitate these enhancements, we can leverage a suite of Hugging Face libraries designed to streamline the implementation of these techniques. The recommended libraries are:

These optimizations are not currently integrated into our repository. However, incorporating them would be highly beneficial, especially for fine-tuning models on large datasets.

For a practical example of how these features could be implemented, consider the following resource: Dreambooth Fine-tuning Colab. This example provides a solid foundation for developing similar capabilities within our projects.

Cruiz102 commented 6 months ago

Accelerate is already added. https://github.com/Cruiz102/Advesarial_Attacks_Tests/commit/fc0c916f8aea426f803d98769a65387142665a4d